Abstract
Observing the actions of others triggers, in our brain, an internal and automatic simulation of its unfolding in time. Here, we investigated whether the instantaneous internal representation of an observed action is modulated by the point of view under which an action is observed and the stimulus type. To this end, we motion captured the elliptical arm movement of a human actor and used these trajectories to animate a photorealistic avatar, a point-light stimulus or a single dot rendered either from an egocentric or an allocentric point of view. Crucially, the underlying physical characteristics of the movement were the same in all conditions. In a representational momentum paradigm, we then asked subjects to report the perceived last position of an observed movement at the moment in which the stimulus was randomly stopped. In all conditions, subjects tended to misremember the last configuration of the observed stimulus as being further forward than the veridical last showed position. This misrepresentation was however significantly smaller for full-body stimuli compared to point-light and single dot displays and it was not modulated by the point of view. It was also smaller when first-person full body stimuli were compared with a stimulus consisting of a solid shape moving with the same physical motion. We interpret these findings as evidence that full-body stimuli elicit a simulation process that is closer to the instantaneous veridical configuration of the observed movements while impoverished displays (both point-light and single-dot) elicit a prediction that is further forward in time. This simulation process seems to be independent from the point of view under which the actions are observed.
Similar content being viewed by others
Introduction
We live in a continuously changing environment and the ability to predict its future states has high behavioral relevance. These predictions are indeed necessary to compensate for the intrinsic delays that our sensory and motor systems have in processing the incoming information and in generating appropriate behavioral responses respectively.
In the specific case of visual perception, its predictive nature was acknowledged very early, already in von Helmholtz’s classic work1 and later confirmed and corroborated by several experimental findings2,3,4,5. A remarkable example of that is representational momentum (RM6), which is “the tendency for observers to misremember the stopping point of an event as being further forward in the direction of movement or change”7.
Representational momentum has been reported for a wide variety of experimental conditions and perceptual domains (see reviews in8,9,10). It was shown to be modulated by several characteristics of the stimulus. For example, it was shown that the degree of forward mislocalization of a moving target depends on both its velocity11,12,13,14 and acceleration14. Furthermore, RM was shown to depend on the degree of perceived “friction” and “gravitational force” experienced by the moving stimulus15,16,17. Finally, for an elliptical motion, it was reported that the displacement of the target reported along the tangential and inward directions were consistent with centripetal forces (18, but see19).
An important class of dynamic stimuli that we encounter in our everyday life is other People’s and our own movements. We are a social species and thus, we constantly engage in social interactions with our conspecifics. At the same time, we also act upon and interact with objects in our environment. As is the case with objects and events in our environment, also interacting with other people and objects relies on simulating how their and our own movements respectively will unfold in time. In agreement with that, several studies have consistently found RM also for stimuli displaying human actions and movements. An early study by Thornton and Hayes reported RM when observing stimuli of crowds moving in different contexts7. Furthermore, recognition performances of briefly presented images of actions are higher when the images are in temporal continuity with previously presented priming movies of those actions20. This effect is robust to changes of the actor in the priming movies and the test images21 and does not occur for biomechanically impossible movements22. Notably, RM has been also reported when human movements were presented by means of impoverished point-light stimuli (PLS), which are stimuli that consist of dots moving with the main joints of an actor performing different actions23. Although the resulting displays consist only of a set of moving dots, they nonetheless convey a compelling impression of a person moving and contain enough information to perform very complex perceptual judgments (for reviews see24,25). In particular, a series of studies using PLS and a temporal occlusion paradigm have shown that our perception continues to internally simulate in real-time an action during an occlusion (i.e. in the absence of the corresponding visual stimulus) with a remarkable degree of precision26,27,28,29.
Despite considerable progress, several questions are still open concerning how we internally represent and simulate observed movements and here we used representational momentum as an experimental tool to address them.
The first important question is whether actions observed from a third- or first-person point of view share similar internal temporal representations. In our everyday life, we observe both others’ actions, from a third-person point of view (3PP), and our own actions, from a first-person perspective (1PP). We know that, in addition to visual areas25,30,31, action observation engages an extended network of parieto-frontal areas (Action Observation Network, AON32,33,34,35,36,37). Interestingly, view-tuned responses have been reported in several areas within this network38,39,40,41,42,43 and several behavioral studies have revealed differences in how our brain processes action stimuli observed from different points of view44,45,46. It is thus conceivable that, actions observed from the first- and third- person view might engage internal simulation mechanisms having different characteristics.
The second important question is to understand how much the internal simulation of observed actions is driven by the characteristics of the motion per se or whether the display or perception of a moving body is necessary. Human movements are governed by a relatively small set of kinematic laws47,48,49,50. Behavioral studies have consistently shown that human perception is biased toward abstract motion stimuli complying with these laws, even when they contain no bodily shape51,52,53,54,55. Furthermore, brain imaging results suggest that observation of these disembodied motion stimuli also produces patterns of neuronal activity that strongly overlap with those produced by action perception56. One might thus speculate that representational momentum reported for human movements might be predominantly due to the kinematic characteristics of human movements and it might not thus require the explicit perception of a human body to be elicited. If, on the other hand, perception of human movements relies on the integration of both form and motion features, as suggested by an influential theoretical account57, then we might expect to observe differences in the RM elicited by normal and impoverished displays.
Addressing these questions poses non-trivial technical problems, as one should be able to show participants exactly the same human movements at different levels of impoverishment and from different points of view. Here, we used modern computer graphics to overcome these problems. Specifically, we motion captured a person during the performance of elliptical trajectories with his right arm. We then used these kinematic data to animate a photorealistic human avatar. We finally rendered the animations from the frontal and first-person perspectives and at different levels of impoverishment, ranging from full-body to point-light and single-dot stimuli. In this manner, we can dissociate the effects on perception of stimuli containing both form and motion features of human movements (full-body), stimuli that contain motion features only and elicit the percept of form features (point-light stimuli) and stimuli that contain only motion features of human movements (single-dot). Notably, all stimuli represented the same underlying physical motion and differed only for the point of view and visual appearance. We then embedded these stimuli into a RM experimental paradigm to address the two questions above.
Methods
Subjects
We performed two Experiments. Twenty-two participants (13 female, age range 22–33 years) and 15 participants (4 female, age range 21–43) participated in Experiment 1 and 2 respectively. All subjects were naïve to the purpose of the experiment, they had normal or corrected-to-normal vision and were compensated for their participation. All procedures were approved by the local Ethics Committees (Comitato Unico Provincia di Ferrara for Experiment 1 and Comitato Etico IRCCS Sicilia Sezione Centro Neurolesi “Bonino-Pulejo” for Experiment 2) and were in accordance with the guidelines of the Declaration of Helsinki. Prior to the experiment, all participants signed an informed consent form.
Stimuli
Stimuli consisted of video clips showing a photorealistic human avatar, point-light stimuli or a moving solid shape. The animations were generated as in58,59. In brief, the movements of a human actor were recorded using a Vicon 612 motion capture system (Vicon Motion Systems Ltd., Oxford, UK) equipped with 9 cameras. The temporal sampling rate was 100 Hz, and the spatial error was less than 1 mm. The recorded movement was a smooth ellipsoidal motion of the right arm along the horizontal plane at a rate of approximately 1 turn per second. The recorded trajectories were low-pass filtered at 5 Hz. A 3 s interval was cropped from the data and imported into commercial software (3D Studio Max and MotionBuilder, both from Autodesk) to animate a commercial avatar model. Animations were then rendered in AVI format at 60 frames per second and at a resolution of 800 × 600 pixels.
We rendered a total of 7 animations showing the movements of an avatar, a point-light stimulus or a single dot as seen from a frontal or first-person point of view (Fig. 1) and the movement of a solid shape (bottom panel of Fig. 5A). The point-light stimulus was generated by parenting, in 3D Studio Max, white spheres to the major joints of the avatar, hiding the avatar body and then rendering the animation from a frontal or first-person of view (middle row in Fig. 1). Similarly, the single-dot stimulus was generated by parenting a white sphere on the tip of the index finger, hiding the avatar body and then rendering the animation of the dot from a first-person or frontal of view (bottom row in Fig. 1). The solid shape stimulus was generated by parenting a “capsule” object to the avatar forearm, hiding the avatar body and then rendering the animation of the moving object from a first-person point of view. Crucially, the physical movements of the joints, dots or solid shape was the same across all animations. The only features that changed across animations were the point of view (first-person or frontal) or the type of stimulus (avatar, point-light display, single dot or solid shape).
Experimental paradigm
The experimental paradigm is shown in Fig. 2. It was implemented in Octave60, running under Linux, using the Psychophysics Toolbox61,62,63. During the experiment, the subjects sat comfortably in front of a 21.5″ commercial LED monitor (Dell U2212HMc, 1920 × 1080 at 60 Hz) at a distance of approximately 60 cm (Experiment 1) or in front of a 17″ LED monitor (Dell Precision M6400) at a distance of approximately 50 cm (Experiment 2). Each trial began with the presentation of one of the six animations. After a random number of frames, uniformly distributed in the interval [50, 80] frames (i.e. [830, 1330] ms), stimulus presentation was interrupted and a black screen was presented for 1 s. After this time interval, a static frame was presented for 1 s. The static frame could have an offset of either − 4, − 2, 0, 2, 4 or 7 frames with respect to the last presented frame of the stimulus, where 0 indicate that it was the same frame, negative numbers indicate that it preceded it in the frame sequence and positive numbers indicate that it followed it in the frame sequence. After 1 s, the static frame disappeared and a message appeared to prompt the subjects to press the key ‘z’ or ‘m’ to indicate whether they perceived the static frame to precede or follow respectively the last presented frame of the animation. The stimulus was centered on the screen along the horizontal axis and it was shifted 50 pixels down from the center along the vertical axis. We used a asymmetric range of temporal offsets (i.e. [− 4, 7]) because, based on previous results in the field (i.e.7,11,17) and prior piloting studies from our lab, we expected a forward shift in time in the subjects’ responses. We thus sampled a larger range of forward displacements to allow for a better fitting of the psychometric function.
Before the experimental session, the subjects were first familiarized with the animations and explained the experimental paradigm. They were then presented with example trials by the experimenter and asked to report their percept verbally. In the example trials, the offset of the static frame was either − 10 or + 10 frames and the experimenter provided verbal feedback. A short practice session lasting few minutes followed. During this session, the offset of the static frame could be either − 5 or + 5 frames and no feedback was provided to the subjects. The practice session could be repeated at the subject’s request and it was followed by the experimental session.
In Experiment 1, during an experimental session, each combination of stimulus type (3 levels: full-body, point-light stimulus and single dot), point of view (2 levels: subjective and frontal view respectively), and offset of the static probe frame (6 levels: − 4, − 2, 0, 2, 4 or 7 frames) was repeated for 15 times for a total of 540 trials. Each subject completed two experimental sessions on different days to collect a total of 30 trials for each combination of stimulus type, point of view and offset of the probe stimulus. Experiment 2 followed the same design, with the only difference that only two stimuli (full-body and solid shape stimuli) and one point of view (first-person) were presented and, given the reduced number of experimental conditions, it was run in a single session. In both experiments, the order of conditions was randomized across subjects (Experiment 1 and 2) and sessions (Experiment 1).
Data analysis
All data analyses were performed in R64. For each subject, we first computed the proportion of ‘before’ responses for each combination of stimulus type, point of view and offset of the static probe frame. The dots in Fig. 3 show the results of this analysis for one of our subjects. For each combination of stimulus type and point of view, we then fitted these results with a psychometric function (see solid curve in Fig. 3) and computed the point of subjective equality (PSE), which is defined as the point where the value of the psychometric function is 0.5 (vertical dotted lines in Fig. 3). The PSE defines the value of the independent variable, which in our experiments was the offset between the static probe frame and the last presented frame, at which the subjects’ perception is perfectly unbiased between the two alternative forced choices. In representational momentum studies, it is used to index the amount of subjects’ perceptual shift65,66,67. Estimation of the PSE was performed by means of the MixedPsy R package68,69 using a generalized linear mixed model and a probit link function. Two participants exhibited patterns of responses that could not be fitted by a psychometric function and were thus excluded from further analysis.
In Experiment 1, we entered the PSEs computed for each subject and condition into a repeated-measures ANOVA, followed, where appropriate, by paired t-tests to further investigate effects of interest. Results of the t-tests were Holm-corrected70 for multiple comparisons. In Experiment 2, since only two conditions were present, we compared the PSEs by means of a paired t-test.
Results
In Experiment 1, we compared PSEs computed across all subjects by means of a two-level repeated-measures ANOVA with factors point of view (2 levels: first-person and frontal view) and stimulus type (3 levels: full-body, point-light and single-dot). Results revealed a significant main effect of the factor stimulus type (F(2,38) = 4.52, p = 0.017), no effect of the factor point of view (F(1, 19) = 0.529, p = 0.48) and no interaction (F(2,38) = 1.76, p = 0.19). To further explore the main effect of the factor point of view (POV), we collapsed behavioral results across the factor point of view (Fig. 4. Results for all factors separately are shown in Fig. S1) and performed paired t-tests between the 3 levels of the factor stimulus type. This analysis revealed a significant difference between the conditions full-body and point-light (p = 0.048, corrected for multiple comparisons) and between the conditions full-body and single-dot (p = 0.037, corrected for multiple comparisons). No significant difference was observed between the conditions point-light and single dot (p = 0.759, corrected for multiple comparisons). This pattern of results suggests that the amount of RM produced by full-body stimuli is significantly smaller for full-body compared to point-light and single-dot displays. All conditions produced a positive RM and this perceptual forward time shift seems not to be modulated by whether the stimulus is observed from an egocentric or allocentric point of view.
One possible explanation for the results in Fig. 4 is that they are predominantly due to the “richness” of the stimulus. In particular, full-body stimuli are perceptually more “salient” as they contain both form and motion information. Explicit form information is instead completely absent in both point-light and single-dot stimuli where it can only be indirectly inferred from the motion of the dots57. It might thus be argued that the additional low-level cues present in full-body stimuli allowed for a more veridical estimation and/or memorization of the last perceived stimulus configuration before disappearance and/or its better comparison with the probe image. To control for this possibility, in Experiment 2, we compared RM produced by full-body stimuli to that of a solid shape having the same motion and orientation of the avatar’s forearm (Fig. 5A, bottom panel). This stimulus possesses relevant low-level cues present in the full-body condition (i.e. it is “solid” and its color is determined by physically veridical illumination), while not resembling a biological effector. Given the non-significance of the factor point of view in Experiment 1, in Experiment 2 we presented only first-person stimuli. Results in Fig. 5 show that a moving solid shape produced significantly higher RM than full-body stimuli (t(14) = − 2.8, p = 0.014). This result suggests that the shorter RM observed for full-body stimuli cannot be explained by the richness of the visual stimulus, per se. Indeed, even the significantly visually richer solid-object stimulus still produced higher RM compared to full-body displays.
Discussion
In this study we used representational momentum (RM6) to investigate the amount of forward shift in time of human perception when observing human bodily movements. In particular, we explored how this shift is modulated by two factors: the type of stimulus (full-body, point-light and single dot) in which the observed movement is embedded and the point of view (first-person or frontal view) from which it is observed. Crucially, we used high-quality computer-generated stimuli to ensure that the same underlying physical motion was presented in all conditions. The main result of our study is that observation of human bodily movements under naturalistic conditions (i.e., a full-body photorealistic avatar) seems to produce a significantly smaller forward shift in time of the observer’s perception compared to when the same physical motion is observed in impoverished displays (point-light stimulus or single-dot display). We found no effect of the point of view. These results provide new insights into how our brain perceives human movements.
The first insight is that the degree of RM elicited by human kinematics depends on the stimulus in which it is embedded. Specifically, full-body stimuli elicited a smaller, but still significantly positive, RM compared to point-light and single-dot stimuli having the same motion characteristics. The subjects’ task was to hold in memory the last perceived stimulus configuration. Thus, a smaller RM for full-body stimuli means that, for this type of stimulus, the memorized configuration was closer to the veridical one at the moment when the movie stimulus was stopped. It is important to emphasize that point-light displays, although not explicitly depicting a human body, elicit a robust and vivid perception of a moving person23,71,72,73,74,75. Thus, our results seem to suggest a “gradient” in how precisely we internally simulate observed actions, with full-body stimuli being more “veridically” simulated than point-light displays. Our result of no difference between point-light and single-dot displays seem to suggest that the former, although also producing a vivid perception of a moving body, engages prediction mechanisms that are more similar to those that would be engaged by a “disembodied” (i.e., containing or eliciting no perception of a moving body) motion.
As mentioned in the Introduction section, observation of full-body and point-light stimuli produces different response patterns at the neuronal level. That is, full-body stimuli elicit robust responses both in temporo-occipital visual areas and in parieto-frontal visuo-motor areas of the action observation network (AON; see for example76). On the contrary, point-light displays seem to engage mostly visual and cerebellar areas25,57,77,78,79,80,81, with only few studies revealing activations within the AON79,82. In previous studies, we showed that while both full-body and abstract dots displays complying with normal human kinematics produce activation within the AON56,58 only for full-body displays these activation patterns are time-locked to the velocity profile of the movement59. Consideration of these results might provide an explanation for the smaller RM produced by full-body compared to single-dot displays reported here. Indeed, we can hypothesize that, in our experiments, the widespread and time-locked activation produced within the AON by the full-body stimuli, when compared to the point-light and single-dot displays, allowed a temporally more veridical internal simulation of the observed action. This more precise internal simulation might be the reason behind the observation of a more accurate estimation of the stimulus configuration at the time when the movie stimulus was stopped.
The second insight provided by our study, is that our internal simulation process does not seem to be tuned for the point of view. Indeed, in our experiments, the factor point of view neither produced a main effect nor was involved in any interaction. In particular, we found that both the first and the third person point of view produced a forward temporal shift in judging the state of the stimulus at the time of its disappearance. This pattern of result is in accordance with a RM study by Brattan et al.67. In their experiments, Brattan et al. asked subjects to judge the correct continuation of a human action after an occlusion. In agreement with our results, they found a positive forward prediction bias, both for actions presented from the first- and the third-person points of view. This bias was not statistically different between the two perspectives. Our study extends this result to point-light and single-dot stimuli, thus corroborating the hypothesis that actions observed from different points of view produce internal simulation processes with similar temporal dynamics. Our results do not however imply that this simulation process is a-modal. There is, on the contrary, experimental evidence suggesting that the point of view is an integral component of how we internally represent and simulate actions. For example, Jarraya et al. found that viewpoint changes interfered with the detection of motion discontinuity in point-light displays of human movements. In a similar vein, Campanella et al.66 found that the discrimination of the size of an unseen object from the observed hand configuration was better in egocentric compared to allocentric view44. Finally, previous neurophysiological studies from our lab, showed that a large percentages of mirror neurons33,37,83,84,85, one of the putative neuronal substrates for the simulation of others’ actions, exhibit view-tuned responses38,39. Taken together, these results and the present study suggest that while several aspects of how we internally represent observed actions are view-dependent, how we internally simulate their unfolding in time appears instead to be view-independent.
The third insight concerns the amount of forward shift in time of the instantaneous internal representation of an internal action. In our experiments, we found that this shift was, on average, 1.15 and 1.03 frames in Experiment 1 and 2 respectively for full-body stimuli, 1.59, 1.64 and 1.71 frames for point-light, single-dot and solid shape displays, respectively. These values correspond to 17 ms and 19 ms (full-body, Experiment 1 and 2 respectively), 26 ms (point-light), 27 ms (single-dot) and 29 ms (solid shape) at the rate of 60 frames per second that we used in our experiments. They are in line with previous experiments67,86 and they further suggest that our perception of observed actions has a remarkable degree of temporal precision.
A potential additional explanation for the lower RM in the full-body condition could be that the richer content of the visual stimulus in this condition, containing both form and motion features, provided more information to the perceptual comparison of the memorized last frame of the movie and the probe stimulus. Results of Experiment 2, showing that RM produced by a solid object was significantly greater than that produced by full-body stimuli, suggest that the “richness” of the stimulus, per se, cannot explain our results. Of course, our control experiment could not necessarily investigate all potential visual features that might modulate RM and future experiments are needed to fully explore this point.
It is worth emphasizing that, to our best knowledge, this is the first study that quantitatively compares RM elicited by full-body, point-light and single-dot stimuli displaying the same underlying physical motion and using the same experimental paradigm. Previous studies used either full-body or point-light stimuli and different experimental paradigms (e.g.67,87). These methodological differences prevented a direct comparison of experimental results. Here, we used instead the same paradigm for all stimuli types and we were thus able to directly compare RM across conditions. This comparison generated several key insights as discussed above. Notably, our approach to leverage modern computer graphics to render the same human movements at different levels of impoverishment can be used to probe additional characteristics of RM during the observation of human movements. For example, one interesting next step will be to investigate whether RM of full-body and impoverished stimuli is modulated by compliance or not of the stimuli with normal human kinematics. From previous studies we know that displays of human movements and abstract motion stimuli activates different brain networks depending on whether the stimuli comply or not with kinematic invariants of human movements56,58. It would thus be interesting to test whether this difference at the neuronal level translates to a difference in RM at the behavioral level.
In conclusion, our results provide evidence that our perceptual mechanisms for the internal simulation of observed actions seem to be differentially engaged by different types of stimulus displays. When the stimulus is a full-body display, our instantaneous internal representation of the stimulus configuration is significantly closer to its veridical physical state than when the stimulus is a point-light, single-dot or solid shape display. Furthermore, our internal simulation mechanisms do not seem to be modulated by the point of view under which the action is observed.
Data availability
Upon publication, the corresponding author (AC) will make the data collected in this study available on an open database. He will also send them directly upon request.
References
von Helmholtz, H. Handbuch der physiologischen Optik. ((English Trans: Southall JPC, Ed), Dover, New York, 1860).
Rao, R. P. & Ballard, D. H. Predictive coding in the visual cortex: A functional interpretation of some extra-classical receptive-field effects. Nat. Neurosci. 2, 79–87 (1999).
Enns, J. T. & Lleras, A. What’s next? New evidence for prediction in human vision. Trends Cogn. Sci. 12, 327–333 (2008).
Nijhawan, R. Motion extrapolation in catching. Nature 370, 256–257 (1994).
Trapp, S. & Bar, M. Prediction, context, and competition in visual recognition. Ann. N. Y. Acad. Sci. 1339, 190–198 (2015).
Freyd, J. J. & Finke, R. A. Representational momentum. J. Exp. Psychol. Learn. Mem. Cogn. 10, 126–132 (1984).
Thornton, I. M. & Hayes, A. E. Anticipating action in complex scenes. Vis. cogn. 11, 341–370 (2004).
Hubbard, T. L. Representational momentum and related displacements in spatial memory: A review of the findings. Psychon. Bull. Rev. 12, 822–851 (2005).
Hubbard, T. L. Toward a general theory of momentum-like effects. Behav. Processes 141, 50–66 (2017).
Thornton, I. M. & Hubbard, T. L. Representational momentum: New findings, new directions. Vis. cogn. 9, 1–7 (2002).
Freyd, J. J. & Finke, R. A. A velocity effect for representational momentum. Bull. Psychon. Soc. 23, 443–446 (1985).
Hubbard, T. L. Cognitive representation of linear motion: Possible direction and gravity effects in judged displacement. Mem. Cognit. 18, 299–309 (1990).
Hubbard, T. L. & Bharucha, J. J. Judged displacement in apparent vertical and horizontal motion. Percept. Psychophys. 44, 211–221 (1988).
Finke, R. A., Freyd, J. J. & Shyi, G. C. W. Implied velocity and acceleration induce transformations of visual memory. J. Exp. Psychol. Gen. 115, 175–188 (1986).
Hubbard, T. L. Cognitive representation of motion: Evidence for friction and gravity analogues. J. Exp. Psychol. Learn. Mem. Cogn. 21, 241–254 (1995).
Hubbard, T. L. Some effects of representational friction, target size, and memory averaging on memory for vertically moving targets. Can. J. Exp. Psychol. Can. Psychol. Exp. 52, 44–49 (1998).
Nagai, M., Kazai, K. & Yagi, A. Larger forward memory displacement in the direction of gravity. Vis. cogn. 9, 28–40 (2002).
Hubbard, T. L. Representational momentum, centripetal force, and curvilinear impetus. J. Exp. Psychol. Learn. Mem. Cogn. 22, 1049–1060 (1996).
Kerzel, D. Centripetal force draws the eyes, not memory of the target, toward the center. J. Exp. Psychol. Learn. Mem. Cogn. 29, 458–466 (2003).
Verfaillie, K. & Daems, A. Representing and anticipating human actions in vision. Vis. cogn. 9, 217–232 (2002).
Verfaillie, K. & Daems, A. Flexible orientation tuning of visual representations of human body postures: Evidence from long-term priming. Front. Psychol. 11, 1–17 (2020).
Kourtzi, Z. & Shiffrar, M. Dynamic representations of human body movement. Perception 28, 49–62 (1999).
Johansson, G. Visual perception of biological motion and a model for its analysis. Percept. Psychophys. 14, 201–211 (1973).
Blake, R. & Shiffrar, M. Perception of human motion. Annu. Rev. Psychol. 58, 47–73 (2007).
Pavlova, M. Biological motion processing as a hallmark of social cognition. Cereb. Cortex 22, 981–995 (2012).
Graf, M. et al. Predicting point-light actions in real-time. Neuroimage 36(Suppl 2), T22-32 (2007).
Parkinson, J., Springer, A. & Prinz, W. Before, during and after you disappear: Aspects of timing and dynamic updating of the real-time action simulation of human motions. Psychol. Res. 76, 421–433 (2012).
Parkinson, J., Springer, A. & Prinz, W. Can you see me in the snow? Action simulation aids the detection of visually degraded human motion. Q. J. Exp. Psychol. 64, 1463–1472 (2011).
Springer, A., Parkinson, J. & Prinz, W. Action simulation: Time course and representational mechanisms. Front. Psychol. 4, 1–20 (2013).
Blakemore, S. J. & Decety, J. From the perception of action to the understanding of intention. Nat. Rev. Neurosci. 2, 561–567 (2001).
Yovel, G. & O’Toole, A. J. Recognizing people in motion. Trends Cogn. Sci. 20, 383–395 (2016).
Casile, A. Mirror neurons (and beyond) in the macaque brain: An overview of 20 years of research. Neurosci. Lett. 540, 3–14 (2013).
Casile, A. Mirror neurons. In Encyclopedia of Behavioral Neuroscience (ed. Della Sala, S.) 541–552 (Elsevier, 2022). https://doi.org/10.1016/B978-0-12-819641-0.00130-4.
Casile, A., Caggiano, V. & Ferrari, P. F. The mirror neuron system: A fresh view. Neuroscience 17, 524–538 (2011).
Keysers, C. & Perrett, D. I. Demystifying social cognition: A Hebbian perspective. Trends Cogn. Sci. 8, 501–507 (2004).
Rizzolatti, G. & Craighero, L. The mirror-neuron system. Annu. Rev. Neurosci. 27, 169–192 (2004).
Rizzolatti, G. & Sinigaglia, C. The mirror mechanism: A basic principle of brain function. Nat. Rev. Neurosci. 17, 757–765 (2016).
Caggiano, V. et al. View-based encoding of actions in mirror neurons of area F5 in macaque premotor cortex. Curr. Biol. 21, 144–148 (2011).
Caggiano, V., Giese, M., Thier, P. & Casile, A. Encoding of point of view during action observation in the local field potentials of macaque area F5. Eur. J. Neurosci. 41, 466–476 (2015).
Maeda, F., Kleiner-Fisman, G. & Pascual-Leone, A. Motor facilitation while observing hand actions: Specificity of the effect and role of observer’s orientation. J. Neurophysiol. 87, 1329–1335 (2002).
Oosterhof, N. N., Tipper, S. P. & Downing, P. E. Viewpoint (in)dependence of action representations: An MVPA study. J. Cogn. Neurosci. 24, 975–989 (2012).
Angelini, M. et al. Perspective-dependent reactivity of sensorimotor mu rhythm in alpha and beta ranges during action observation: An EEG study. Sci. Rep. 8, 1–11 (2018).
Alaerts, K., Heremans, E., Swinnen, S. P. & Wenderoth, N. How are observed actions mapped to the observer’s motor system? Influence of posture and perspective. Neuropsychologia 47, 415–422 (2009).
Campanella, F., Sandini, G. & Morrone, M. C. Visual information gleaned by observing grasping movement in allocentric and egocentric perspectives. Proc. R. Soc. London. Part B 278, 2142–2149 (2011).
Watanabe, R. & Higuchi, T. Behavioral advantages of the first-person perspective model for imitation. Front. Psychol. 7, 1–8 (2016).
Tidoni, E., Scandola, M., Orvalho, V. & Candidi, M. Apparent biological motion in first and third person perspective. Iperception 7, 1–6 (2016).
Flash, T. & Hochner, B. Motor primitives in vertebrates and invertebrates. Curr. Opin. Neurobiol. 15, 660–666 (2005).
Viviani, P. & Terzuolo, C. Trajectory determines movement dynamics. Neuroscience 7, 431–437 (1982).
Lacquaniti, F., Terzuolo, C. & Viviani, P. The law relating the kinematic and figural aspects of drawing movements. Acta Psychol. 54, 115–130 (1983).
Flash, T. & Hogan, N. The coordination of arm movements: an experimentally confirmed mathematical model. J. Neurosci. 5, 1688–1703 (1985).
Viviani, P. & Stucchi, N. Biological movements look uniform: Evidence of motor-perceptual interactions. J. Exp. Psychol. Hum. Percept. Perform. 18, 603–623 (1992).
Viviani, P. & Stucchi, N. The effect of movement velocity on form perception: Geometric illusions in dynamic displays. Percept. Psychophys. 46, 266–274 (1989).
Bidet-ildei, C., Orliaguet, J.-P., Sokolov, A. N. & Pavlova, M. Perception of elliptic biological motion. Perception 35, 1137–1147 (2006).
Flach, R., Knoblich, G. & Prinz, W. The two-thirds power law in motion perception. Vis. Cogn. 11, 461–481 (2004).
Levit-Binnun, N., Schechtman, E. & Flash, T. On the similarities between the perception and production of elliptical trajectories. Exp. Brain Res. 172, 533–555 (2006).
Dayan, E. et al. Neural representations of kinematic laws of motion: Evidence for action-perception coupling. Proc. Natl. Acad. Sci. USA 104, 20582–20587 (2007).
Giese, M. A. & Poggio, T. Neural mechanisms for the recognition of biological movements. Nat. Rev. Neurosci. 4, 179–192 (2003).
Casile, A. et al. Neuronal encoding of human kinematic invariants during action observation. Cereb. Cortex 20, 1647–1655 (2010).
Agosta, S., Battelli, L. & Casile, A. Human movements and abstract motion displays activate different processes in the observer’s motor system. Neuroimage 130, 184–193 (2016).
Eaton, J. W., Bateman, D., Hauberg, S. & Wehbring, R. GNU Octave version 5.2.0 manual: A high-level interactive language for numerical computations. https://www.gnu.org/software/octave/doc/v5.2.0/ (2020).
Kleiner, M., Brainard, D. H. & Pelli, D. G. What’s new in Psychtoolbox-3? In 30th European Conference on Visual Perception (ECVP) (2007).
Brainard, D. H. The psychophysics toolbox. Spat. Vis. 10, 433–436 (1997).
Pelli, D. G. The VideoToolbox software for visual psychophysics: Transforming numbers into movies. Spat. Vis. 10, 437–442 (1997).
R Core Team. R: A language and environment for statistical computing. https://www.r-project.org/ (2021).
Blättler, C., Ferrari, V., Didierjean, A., van Elslande, P. & Marmèche, E. Can expertise modulate representational momentum? Vis. Cogn. 18, 1253–1273 (2010).
Jarraya, M., Amorim, M.-A. & Bardy, B. G. Optical flow and viewpoint change modulate the perception and memorization of complex motion. Percept. Psychophys. 67, 951–961 (2005).
Brattan, V. C., Baker, D. H. & Tipper, S. P. Spatiotemporal judgments of observed actions: Contrasts between first- and third-Person perspectives after motor priming. J. Exp. Psychol. Hum. Percept. Perform. 41, 1236 (2015).
Moscatelli, A., Balestrucci, P. Psychophysics with R: The R package MixedPsy. (2017).
Moscatelli, A., Mezzetti, M. & Lacquaniti, F. Modeling psychophysical data at the population-level: The generalized linear mixed model. J. Vis. 12, 26–26 (2012).
Holm, S. A simple sequentially rejective multiple test procedure. Scand. J. Stat. 6, 65–70 (1979).
Troje, N. F., Westhoff, C. & Lavrov, M. Person identification from biological motion: Effects of structural and kinematic cues. Percept. Psychophys. 67, 667–675 (2005).
Casile, A. & Giese, M. Critical features for biological motion. J. Vis. 3, 85 (2003).
Thornton, I. M., Pinto, J. & Shiffrar, M. The visual perception of human locomotion. Cogn. Neuropsychol. 15, 535–552 (1998).
Neri, P., Morrone, M. C. & Burr, D. C. Seeing biological motion. Nature 395, 349–351 (1998).
Pavlova, M. & Sokolov, A. Orientation specificity in biological motion perception. Percept. Psychophys. 62, 889–899 (2000).
Molenberghs, P., Cunnington, R. & Mattingley, J. B. Brain regions with mirror properties: A meta-analysis of 125 human fMRI studies. Neurosci. Biobehav. Rev. 36, 341–349 (2012).
Jastorff, J. & Orban, G. A. Human functional magnetic resonance imaging reveals separation and integration of shape and motion cues in biological motion processing. J. Neurosci. 29, 7315–7329 (2009).
Vaina, L. M., Solomon, J., Chowdhury, S., Sinha, P. & Belliveau, J. W. Functional neuroanatomy of biological motion perception in humans. Proc. Natl. Acad. Sci. 98, 11656 (2001).
Saygin, A. P., Wilson, S. M., Hagler, D. J., Bates, E. & Sereno, M. I. Point-light biological motion perception activates human premotor cortex. J. Neurosci. 24, 6181–6188 (2004).
Sokolov, A. A., Erb, M., Grodd, W. & Pavlova, M. A. Structural loop between the cerebellum and the superior temporal sulcus: Evidence from diffusion tensor imaging. Cereb. Cortex https://doi.org/10.1093/cercor/bhs346 (2012).
Sokolov, A. A. et al. Biological motion processing: The left cerebellum communicates with the right superior temporal sulcus. Neuroimage 59, 2824–2830 (2012).
Jastorff, J., Abdollahi, R. O., Fasano, F. & Orban, G. A. Seeing biological actions in 3D: An fMRI study. Hum. Brain Mapp. 37, 203–219 (2016).
Gallese, V., Fadiga, L., Fogassi, L. & Rizzolatti, G. Action recognition in the premotor cortex. Brain 119, 593–609 (1996).
Rizzolatti, G., Cattaneo, L., Fabbri-Destro, M. & Rozzi, S. Cortical mechanisms underlying the organization of goal-directed actions and mirror neuron-based action understanding. Physiol. Rev. 94, 655–706 (2014).
Bonini, L., Rotunno, C., Arcuri, E. & Gallese, V. Mirror neurons 30 years later: Implications and applications. Trends Cognit. Sci. 26, 767–781. https://doi.org/10.1016/j.tics.2022.06.003 (2022).
Craighero, L. et al. Temporal prediction of touch instant during observation of human and robot grasping. Brain Res. Bull. 75, 770–774 (2008).
Sparenberg, P., Springer, A. & Prinz, W. Predicting others’ actions: Evidence for a constant time delay in action simulation. Psychol. Res. 76, 41–49 (2012).
Acknowledgements
The authors acknowledge the support of the APC central fund of the University of Messina. We thank Prof. Laila Craighero for insightful discussions during the preparation of this manuscript.
Author information
Authors and Affiliations
Contributions
A.C. and E.Z. conceived the study. E.Z. and D.B. performed the study. A.C. and D.B. analyzed the data. A.C. drafted the initial manuscript and D.B. its revised versions. A.C., D.B. and E.Z. finalized the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Zucchini, E., Borzelli, D. & Casile, A. Representational momentum of biological motion in full-body, point-light and single-dot displays. Sci Rep 13, 10488 (2023). https://doi.org/10.1038/s41598-023-36870-2
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-023-36870-2
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.