Abstract
Seeing an object is a natural source for learning about the object’s configuration. We show that language can also shape our knowledge about visual objects. We investigated sign language that enables deaf individuals to communicate through hand movements with as much expressive power as any other natural language. A few signs represent objects in a specific orientation. Sign-language users (signers) recognized visual objects faster when oriented as in the sign, and this match in orientation elicited specific brain responses in signers, as measured by event-related potentials (ERPs). Further analyses suggested that signers’ responsiveness to object orientation derived from changes in the visual object representations induced by the signs. Our results also show that language facilitates discrimination between objects of the same kind (e.g., different cars), an effect never reported before with spoken languages. By focusing on sign language we could better characterize the impact of language (a uniquely human ability) on object visual processing.
Similar content being viewed by others
Introduction
Much of what humans know about visual objects is learned implicitly by seeing the objects and is shaped by their sensibility for recognizing object features that are encountered more commonly and predictably. The brain mechanisms supporting this type of learning are largely shared with other animal species1. But humans evolved the unique ability to create and manipulate symbols that represent visual objects and are understood to stand for the objects themselves. Do these symbols provide humans with an additional source for learning implicitly about visual objects? Could a symbol-based form of knowledge enrich what is learned directly from visual objects? The spoken languages that emerged from human evolution are built on words, symbols providing humans a powerful cognitive tool for identifying visual objects. In light of their representative capacity, words would stand as primary candidates for contributing to knowledge about visual objects. Words have consistently been found to modulate visual discrimination, affecting the initial stages of color and object identification2,3,4. These word effects have been linked to words’ primary role in defining conceptual categories3,5. As category labels, words participate in strengthening conceptual boundaries and grouping similar objects. Words can thus down-regulate even early processes of visual object recognition by providing diagnostic information for disentangling different sources of neural activation.
Distinguishing individual exemplars requires identifying variations in the features that set apart each visual object within its category. It is mainly by linking words into phrases (“The red car”; “The long car”) that speakers can mark these distinctions. Individual words are in this respect quite ineffectual given their lack of reference to specific visual features – nothing in the form of the word car tells us of the shape, orientation or size of a given car. This limitation is not shared by signs, the equivalent of words in the sign languages that deaf people use to communicate. As signing is grounded in shaping and moving the hands and in changes of facial expression6,7,8, signs can display specific visual features of the corresponding objects, including shape, orientation, and size9,10,11. We turned to signs to investigate if linguistic symbols provide the kind of knowledge needed to discriminate between different instances of an object, which relates to object features such as orientation, color, size, and shape.
Studies have shown that knowing a sign language improved performance in a variety of visuo-spatial tasks, which in turn led to the realization that signing affects spatial cognition12, response mapping13, visual mental imagery14, and face processing15. The effects of sign language have been associated with specific characteristics of the language, including the use of hand configuration, location, and facial expression to mark grammatical and lexical distinctions. For example, signers’ ability to rotate visual images was related to the use of sign language to describe spatial relations, which signers define from a different perspective than the viewers who try to understand them14. The effects of sign language documented in prior studies were general in nature, stemming from core features of the language and implicating visual processes that, like mental imagery, apply to a wide range of stimuli. We explored if the effects of sign language are more specific, involving distinctive object features that affect the appearance and identity of objects from a given category. We focused on object orientation. Orientation is a diagnostic feature of object variation – changes in object orientation are associated with variations in object’s position and motion or reflect interactions with other objects (e.g., hitting). Changes in orientation also help to recognize if there are multiple instances of the same objects. By investigating orientation, we could examine effects related to the recognition of specific objects in a given category.
As illustrated in Fig. 1, signs from multiple languages iconically display objects in a specific orientation, as with the sign fish, showing fish with the head pointing toward the signer’s left side. We took advantage of this distinct aspect of signs to examine if linguistic symbols enable within-category discriminations. In Experiment 1, signers and speakers decided if words matched the pictured objects. Object orientation either aligned with or differed from the object orientation embedded in the sign (Fig. 2a). If signs shape expectations about object orientation, only signers’ responses should be affected by whether signs displayed objects in a particular orientation. Previous studies reported faster responses in a sign-picture matching task when the visual features of the objects shown in the pictures were iconically represented in the signs16. In contrast with previous studies, our task did not involve sign presentation and was not based on the perceived similarity of visual stimuli.
In Experiment 2 orientation was similarly varied, where we recorded the brain’s event-related potentials (ERPs) elicited by changes in object orientation. We anticipated that only signers’ ERPs would show brain correlates associated with the explicit encoding of orientation by the sign.
Experiment 1
We selected 28 objects, each represented by a sign displaying the object in a specific orientation. The example of screwdriver is presented in Fig. 2. As illustrated in Fig. 2, the orientation in which an object appears in the sign varies between signer and viewer perspectives. As we did not know which perspective was critical, we tested both of them. Each object appeared in two sign-consistent orientations (signer- and viewer-centered) and a sign-inconsistent orientation (Fig. 2a). To test deaf signers (n = 24) and speakers (n = 24) with the same task, we presented written words. Thus, participants decided as fast as possible whether or not the written word matched the picture presented next. In half the trials the word and picture matched (Yes responses), in the other half they did not (No responses). We expect picture orientation to affect Yes responses for signers but not for speakers, who are unaware of sign language. We analyzed the effects of the predictors Orientation (signer-centered vs. viewer-centered vs. signer-inconsistent) and Group (signers vs. speakers) on the speed of Yes responses.
The results are displayed in Fig. 2b. Statistical analyses showed no effects of Orientation or Group (signers vs. speakers), ps > 0.18. Critically, the interaction between these two variables was significant, χ2 = 7.43, p = 0.024. Pairwise comparisons (fdr corrected) indicated that deaf participants responded faster to the signer-centered orientation and the viewer-centered orientation compared to the sign-inconsistent orientation, z = 2.97, p = 0.008 and z = 2.50, p = 0.018, respectively. With speakers these comparisons failed to reach the significance level, ps > 0.72. In essence, signers demonstrated a distinct responsiveness to orientation traceable to their knowledge of signs.
Experiment 2
Using a semantic judgment task, Meade et al.17 showed automatic sign activation in bilingual signers during word reading. Experiment 2 aimed to investigate the orientation effect without presenting words. We used ERPs to shed light on the brain underpinnings of the orientation effect found in the previous experiment, specifically to determine whether it involved object perceptual processing or response-related mechanisms. The ERP technique offers the unique opportunity to investigate the time course of cognitive processes even in trials in which no response is required. In Experiment 2 we measured brain’s event-related potentials (ERPs) elicited by orientation change in a three-stimuli oddball task. All the stimuli were object drawings. The task required responding to an infrequent target shown in a background of two other objects, one occurring frequently (standard) the other rarely (deviant). No responses are required to standard and deviant stimuli (Fig. 3). Deviants typically elicit larger positive amplitude of the electric signal compared to standards18,19,20,21,22. Furthermore, amplitude differences were found between standards and deviants varying in orientation, a result proving the paradigm’s sensitivity to object orientation23. In Experiment 2, the picture of a pen appeared as standard and deviant with an orientation that either aligned with the orientation encoded in the sign pen, from the signer perspectives, or with an orientation that differed from it (Fig. 3). In line with the results of Experiment 1, the orientation of pen was expected to elicit different responses in signers and speakers. To control that the effects found with signers for pen stemmed from the orientation in which pen is displayed in the sign, two additional conditions were included in Experiment 2. We presented the picture of a chair and a bar similar in shape to pen. Both stimuli were presented as standards and deviants, varying their orientation (Fig. 3). Since the sign chair does not encode object orientation, we expected the effects found with pen not to replicate with chair. Bar was shown in the same orientation as pen. We expected bar not to elicit orientation effects analogous to pen.
The oddball task involves discriminating if information about the incoming stimulus matches the information previously acquired about the standard18,19,20,21,22,23. This process is indexed by the deviancy effect, which arises over the posterior regions around 150 ms after stimulus onset, surfacing later at 250–500 ms as a positive amplitude increase over the midline electrodes, the typical configuration of P300 wave. When different colors or objects were shown as standard and deviant2,3, larger deviancy effects appeared at 150–200 ms in the languages where distinct words existed for the stimuli – e.g., in English, where the cup/mug distinction is lexically marked, but not in Spanish where the same word is used for both objects. As category labels, words thus appear to affect the earlier processes supporting initial, crude object discriminations3,5. The within-category discrimination required in the present study to distinguishing between standards and deviants, would depend on more refined visual object representations that are accessed at later stages or processing18. We therefore expect that the later P300 component would index the effects of signs on orientation.
In a first data-driven analysis, we compared the deviancy effect elicited in signers and speakers by each object (pen, chair, bar) from stimulus onset up to 600 ms, using non-parametric cluster-based permutations. A larger deviancy effect for signers was found only for pen between 380 and 500 ms after stimulus presentation over central/parietal electrodes, p = 0.025, (Fig. 4a). These results aligned with our predictions of group differences limited to pen and the P300 component. Next, we carried out planned analyses on P300 amplitude, examining the effects of the variables Deviancy (Standard vs. Deviant), Group (Signers vs. Speakers), and Stimulus Type (Pen vs. Chair vs. Bar) on the responses recorded at 250–550 ms over the Pz electrode. The significant triple interaction revealed that the deviancy effect varied between groups and across stimuli, F(2, 36) = 5.47, p = 0.008, ηG2 = 0.036 (Fig. 4b). Additional analyses were carried out to elucidate the nature of the interaction. First, when Deviancy and Stimulus Type were examined separately for signers and speakers, the interaction was significant only for signers, F(2, 18) = 7.44, p = 0.004, ηG2 = 0.116; for speakers, F < 1. Second, Deviancy and Group were examined separately for each type of stimuli; a significant interaction was found only for pen, F(1, 18) = 7.72, p = 0.012, ηG2 = 0.121; other Fs <1 (Fig. 5a). These additional analyses showed that in P300, differences between signers and speakers are confined to pen. A final analysis aimed to determine if the effects related to pen were driven by the standard or the deviant. The variables Group and Distractor Type were examined separately for standard and deviant. The interaction was significant for the deviant and not for the standard, F(2, 36) = 5.79, p = 0.007, ηG2 = 0.087; F < 1. The latter finding revealed that group differences were elicited only by the deviant (Fig. 5b).
General Discussion
The orientation effects we found were specific, first for appearing with signers and not speakers (Exp. 1 and 2), and second for involving only objects whose signs visually encode object orientation (Exp. 2). This specificity was expected under the hypothesis that signers are especially responsive to objects oriented as in the signs. This hypothesis also anticipated – as we found – that the orientation effects would engage the brain mechanisms of object processing indexed by P300.
Importantly, the specificity of our findings ruled out that brain processes other than those supporting visual object recognition underlay the sign effect. Action processing would be a plausible candidate considering the relevance of object orientation for actions toward the object (e.g., grasping) along with findings showing that even viewing an object picture could trigger the planning of an action24,25,26,27. However, an explanation in terms of action processing is inconsistent with the specificity of our findings. As signers and speakers have had reasonably similar experiences in using objects, this explanation anticipates their responses not to differ. Another plausible candidate is mental rotation, possibly needed in our tasks to align the stimuli to a canonical orientation. Notably, differences between signers and speakers could have stemmed from signers’ skillfulness in mental rotation14. However, an explanation in terms of mental rotation predicts group differences with all the objects we tested, not just those with signs embedding object orientation.
In what way would signs affect object orientation? Under a lexical hypothesis, it would result from the implicit and automatic activation of information concerning the sign17. This hypothesis aligns with accounts proposing that higher levels of processing influence lower ones, such as the interactive-activation models28,29. According to this view, bottom up sensory processes and prior knowledge are jointly used and integrated in order to determine the contents of the perceptual experience30. From this perspective, the visual object would activate information about the shape, location, orientation, and movement involved with the sign. The effects found in our tasks would then depend on whether or not objects aligned with sign orientation. Thus, for example, matching responses would be faster in Experiment 1 when the orientation of the object and the sign coincided, and slower otherwise. An a-posteriori test of the lexical hypothesis was conducted on the basis of prior evidence from picture signing. ERPs correlates of lexical activation were found when comparing iconic and non-iconic signs31. Iconic signs, which reproduce distinguishing features of objects, elicited more positive amplitudes than non-iconic signs in posterior sites at around 100 ms post stimulus-onset (P100). We inspected if similar differences arose in Experiment 2 between pen (iconic) and chair (non-iconic) in posterior-occipital regions at 70–140 ms (see Methods for details). We found no iconicity effects, Fs < 1, disconfirming the lexical hypothesis. It should be noted, however, that differences in the paradigms used (oddball paradigm Vs. to standard picture naming) could also account for the lack of iconicity effect in our study and therefore the hypothesis of automatic and implicit sign activation cannot be fully rejected.
Alternatively, signs could affect object representations by changing feature weights, a cognitive correlate of feature distinctiveness shaping expectations. Their larger weights make salient features more expected, so that their detection becomes crucial in object recognition. With orientation, the weight increase would make certain orientations more canonical and therefore more expected. Embodying object orientation into the sign could also strengthen the saliency of certain orientations. Therefore, as we expect to see a bottle upright or a monitor facing us32, signers would expect objects to align as in the signs. Individual signs would thus warp visual perception by sensitizing to the object features they transparently encode.
Our findings revealed that by affecting the processing of specific visual features and distinct objects, signs contribute to processes enabling the discrimination of different instances of an object. Signs re-map the boundaries of the language-vision interface drawn upon spoken language. While prior evidence showed that words affect categorical discriminations and modulate earlier neural processes in visual object recognition5,33,34, our results from signs suggest language effects in the later stages of object processing responsible for detecting object changes. Disclosing more fully the language’s potential to shape visual cognition, signs point to other types of symbols incorporating object features that may also affect within-category discriminations. One example is represented by spoken words sounding like noises, calls, object sounds, voice features or/musical elements, as with the English words bang, quack, click, grunt or bong, respectively. These types of words could influence acoustic discrimination. If the reason for the sign effects documented in our experiments is that the complexity of object processing makes any processing aids valuable, human brains should also exploit the opportunities offered by any type of symbol embodying object features35. In fact, permeability to symbols would be a necessity, and taking advantage of the opportunities offered by symbols an adaptive choice.
Methods
Experiment 1
Participants
Required sample was estimated using the software GPower 3.136, based on the effect size obtained by Thompson et al.11. Precisely, Signer participants (n = 24, 13 males; age: mean = 31.3, range = 18–53, SD = 10.1) presented with profound bilateral hearing loss (>90 dB). 16 participants were exposed to Italian Sign Language [Lingua Italiana dei segni, LIS] from birth, acquiring it from deaf parents, and used it as their primary and preferred form of communication; 8 participants acquired LIS between 3–12 years old (mean = 8, SD = 3.3) and had at least 10 years of exposure to LIS (mean = 21.6, range = 10–42, SD = 9.9). Signers were fluent in LIS (mean rate = 6.6, range = 4–7, SD = 0.8, on a 7-point scale from 1-no proficiency, to 7-native signer). Hearing native Italian speakers with no knowledge of sign language were recruited for the control group (n = 24, 10 males; mean age = 22.3, range = 20–29, SD = 2.19). All participants were right-handed and had normal or corrected-to-normal vision. Individuals participated in Experiment 1 and 2 voluntarily and provided informed consent. The study was approved by the Ethical Committee for Psychological Research of the University of Padova (Number 2455). Informed consent was obtained from all participants. Experiments have been performed in accordance with relevant guidelines and regulations.
Materials and procedure
We tested 28 pictured objects (stimulus list in Appendix; pictures available at https://osf.io/y4pb9/). Each pictured object was shown in black-and-white in three different orientations (Fig. 2). In two of these orientations, objects were oriented as in the corresponding signs viewed from the perspective of the signer (signer-centered) or the viewer’s (viewer-centered). Specular images of the same object were created to show these orientations. In a third orientation (sign-inconsistent), objects were rotated ~90° from the signer-centered perspective. An object appeared in a specific orientation twice, once preceded by the object name (Yes response) and once by the name of another object tested in Experiment 1 (No response).
Each experimental trial started with a fixation point displayed for 300 ms and replaced by a written word (Times Roman font, 28 points) presented for 2 s; next, a blank screen appeared for a randomly selected duration of 200, 400, or 800 ms, followed by a picture that stayed in view until the response or up to 3 s. The next trial started automatically after 1 or 2 s. Words and pictures appeared in the center of the computer screen. Participants decided as fast and accurately as possible if words and pictures matched by pressing the s key of the keyboard with the left hand for Yes responses (“Sì” in Italian), and the n key with the right hand for No responses (“No” in Italian). Participants were first familiarized with the task with a novel set of 4 warm-up words and pictures for 8 trials (4 Yes responses, 4 No responses). The 168 trials were presented in a pseudo-randomized order, so that pictures and words were not repeated in the next trial, and the same response did not occur in more than four consecutive trials. There was a short pause after 42 trials. Response times (RTs) were measured from picture onset. Stimuli presentation and response recording were controlled by E-Prime 2.0 (Psychology Software Tools, Inc., Sharpsburg, Pittsburgh, PA).
Analyses
RT-outliers (2.5% of the responses) were removed using the procedure recommended by Van Selst and Jolicoueur37. Incorrect responses occurred rarely with both signers (2.2%) and speakers (3.1%) and were not analyzed. Analyses were performed on Yes responses. Given the slower reading-times typically experienced by deaf readers38,39, raw scores of the RTs of the two groups were transformed into z scores to correct for differences in processing speed and variability across groups of participants40. Z scores were analyzed through linear mixed effects models and accuracies through generalized linear mixed effects models using the lm441 package in R software42. Predictors were Participants (signers, speakers) and Orientation (signer-centered, viewer-centered, sign-inconsistent), and we included by-item and by-participant random effects.
Experiment 2
Participants
Based on Azizian et al.43 who measured the P300 effect in an oddball paradigm with stimuli from the same category and established the expected effect-size, we opted for a sample of 20 participants overall. This choice took into account (a) the difficulty in recruiting right-handed profound deaf signers, aged below 40 years, without cochlear implants, who were exposed to sign language from birth or early childhood, familiar with the same variant of Italian Sign Language used in the city of Padua, agreeing to EEG measurement and (b) the fact that we examined a robust psychophysiological effect. Signers (n = 10, 4 males; age: mean = 21.1, range = 18–32, SD = 4.3) were congenital deaf with profound bilateral hearing loss (>90 dB) who were exposed to LIS from birth, acquired it from deaf parents and used it as their primary and preferred form of communication. Speakers were hearing native Italian speakers with no knowledge of any sign language (n = 10, 4 males, age: mean = 22.9, range = 19–29, SD = 2.7). None of the participants participated in Experiment 1. All participants were right-handed and had normal or corrected-to-normal vision. The study was approved by the Ethical Committee for Psychological Research of the University of Padova (Number 2455). Informed consent was obtained from all participants. Experiments have been performed in accordance with relevant guidelines and regulations.
Materials and procedure
We selected three sets of visual stimuli, each comprising one target, one standard, and one deviant. In each set, standard and deviant showed the same object, but in different orientations (Fig. 3; stimuli are available at https://osf.io/y4pb9/). The stimuli were manipulable objects in two of the sets (spoon and pen; key and chair), and geometrical shapes in the other set (square and bar). Each stimulus was presented in three orientations. Pen was shown in two orientations. Bar was oriented as pen. Chair was rotated either toward the right or the left, the mirror-image rotation that signers perform especially skillfully14. Objects were displayed as black-and-white drawings on a white background, whereas the geometrical shapes were shown in white on a black background. Stimuli subtended approximately 8° of visual angle and were presented in the center of a CRT monitor. Stimuli appeared with different probabilities: 0.80 for standards, 0.15 for deviants, and 0.05 for targets (see Fig. 3). The experiment comprised six blocks of 480 trials each. Blocks were presented in counterbalanced order. The same object appeared as standard and deviant in two consecutive blocks; the object orientation that was assigned to standard and deviant was reversed between these blocks.
Stimuli appeared for 300 ms, separated by inter-stimulus intervals that varied randomly in duration (400, 450, 500, 550 or 600 ms) but were used evenly. Stimuli presentation followed a quasi-random order in which targets and deviants never appeared in consecutive trials and the next deviant was never shown until at least three standards appeared in a row. Instructions, which were written in Italian and presented on the computer screen at the beginning of each block, informed of the next target (key, spoon, or square). Responses to targets were made by pressing the space bar of the keyboard. E-Prime 2.0 (Psychology Software Tools, Inc., Sharpsburg, Pittsburgh, PA) was used for stimuli presentation and response recording. Participants were seated comfortably at 60 cm from the computer screen in a noise- and light-attenuated room. The experimental session lasted about 1.5 hours.
EEG recording and data analysis
EEG activity was recorded continually using 32 Ag-AgCl electrodes that were referenced to the left earlobe and mounted on an elastic cap according to the extended 10–20 system. Horizontal EOG (HEOG) was recorded bipolarly from electrodes positioned lateral to the outer canthi of both eyes. Vertical EOG (VEOG) was recorded bipolarly from two electrodes, one above (Fp1) and one below the left eye. EEG, HEOG, and VEOG signals were amplified and filtered using a bandpass of 0.01–80 Hz and digitized at a sampling rate of 250 Hz. Impedance at each electrode site was maintained below 5 kΩ.
EEG recordings were processed offline using the MATLAB toolbox EEGLAB44. EEG was re-referenced off-line to the average of the left and right earlobes, and low- and high-pass filtered (30-0.1 Hz). The EEG was segmented into 700 ms epochs starting 100 ms prior to the onset of the stimuli. Epochs at each electrode site were baseline-corrected with reference to mean activity within the 100 ms preceding stimuli presentation. Bad channels were interpolated with the surrounding electrodes. With 3 participants one bad channel was interpoled. Epochs were subjected to independent component analysis (ICA45) to detect highly characteristic artefacts (e.g., eye blinks). Artefactual components were discarded, along with epochs that contained excessive noise or drift (signals exceeding ± 100 μV at any electrode). Epochs were averaged for each participant across the two conditions of interest (standards and deviants). Each average was based on data from at least 90 trials.
A data-driven analysis was conducted to examine the unfolding of the deviancy effect (deviant–standard) using non-parametric cluster-based permutations computed on ERP amplitudes from stimulus onset up to 600 ms. Specifically, we examined differences between signers and speakers in the deviancy effect (standard vs. deviant), separately for each standard-deviant pair (pen, chair, bar). Spatial threshold was fixed at 2 clustered electrodes, and we used 2000 randomizations. Analyses were performed using the Brainstorm software package46 with the function FT_timeskects47. Another kind of analysis focused on P300. When (as in our task) target discrimination is easy and demand of focal attention is minimal, the largest P300 amplitude is observed over central/parietal electrodes48. We examined the ERPs differences between standards and deviants over the Pz electrode. Mean amplitudes for standards and deviants were examined using a mixed repeated-measures ANOVA; Greenhouse-Geisser correction for non-sphericity was applied when appropriate. A final analysis, modelled after Baus and Costa31, examined the iconicity effects on signers’ responses to pen (iconic) and chair (non-iconic) at 70–140 ms, in two electrode clusters (Posterior: Pz, P3 and P4; Occipital: Oz, PO9 and PO10). To discard effects of deviancy, this analysis was restricted to standards.
Data availability
The datasets generated during and/or analysed during the current study are available at https://osf.io/y4pb9/?view_only=4357b4a8a46f481f93ea94a3eebd7f9c.
References
Santolin, C. & Saffran, J. R. Constraints on statistical learning across species. Trends Cogn. Sci. 22, 52–63 (2018).
Boutonnet, B., Dering, B., Viñas-Guasch, N. & Thierry, G. Seeing objects through the language glass. J. Cognitive Neurosci. 25, 1702–1710 (2013).
Thierry, G., Athanasopoulos, P., Wiggett, A., Dering, B. & Kuipers, J. R. Unconscious effects of language- specific terminology on preattentive color perception. P. Natl. Acad. Sci. USA 106, 4567–4570 (2009).
Winawer, J. et al. Russian blues reveal effects of language on color discrimination. P. Natl. Acad. Sci. USA 104, 7780–7785 (2007).
Lupyan, G. & Clark, A. Words and the world: Predictive coding and the language-perception-cognition interface. Curr. Dir. Psychol. Sci. 24, 279–284 (2015).
Klima, E. S., & Bellugi, U. The signs of language. Cambridge, MA: Harvard University Press (1979).
Sandler, W., & Lillo-Martin, D. Sign language and linguistic universals. Cambridge, United Kingdom: Cambridge University Press (2006).
Stokoe, W. C. Sign language structure. In Studies in Linguistics: Occasional Papers, Vol. 8. Buffalo, NY: University of Buffalo Press (1960).
Taub, S. F. Language from the body: Iconicity and metaphor in American Sign Language. New York, NY: Cambridge University Press (2001).
Emmorey, K. Iconicity as structure mapping. Phil. Trans. R. Soc. B 369, 20130301 (2014).
Thompson, R. L., Vinson, D. P. & Vigliocco, G. The link between form and meaning in American Sign Language: Lexical processing effects. J. Exp. Psychol. Learn. Mem. Cogn. 35, 550–557 (2009).
Pyers, J. E., Shusterman, A., Senghas, A., Spelke, E. S. & Emmorey, K. Evidence from an emerging sign language reveals that language supports spatial cognition. P. Natl. Acad. Sci. USA 107, 12116–12120 (2010).
Peressotti, F., Scaltritti, M. & Miozzo, M. Can sign language make you better at hand processing? PloS one 13, e0194771 (2018).
Emmorey, K., Kosslyn, S. & Bellugi, U. Visual imagery and visual-spatial language: Enhanced imagery abilities in deaf and hearing ASL signers. Cognition 46, 139–181 (1993).
Stoll, C. et al. Face recognition is shaped by the use of sign language. J. Deaf Stud. Deaf Edu. 23, 62–70 (2017).
Vinson, D., Thompson, R. L., Skinner, R. & Vigliocco, G. A faster path between meaning and form? Iconicity facilitates sign recognition and production in British Sign Language. J. Mem. Lang. 82, 56–85 (2015).
Meade, G., Midgley, K. J., Sehyr, Z. S., Holcomb, P. J. & Emmorey, K. Implicit co-activation of American Sign Language in deaf readers: An ERP study. Brain Lang. 170, 50–61 (2017).
Polich, J. Updating P300: an integrative theory of P3a and P3b. Clin. Neurophysiol. 118, 2128–2148 (2007).
Comerchero, M. D. & Polich, J. P3a and P3b from typical auditory and visual stimuli. Clinical Neurophysiol. 110, 24–30 (1999).
Courchesne, E., Hillyard, S. A. & Galambos, R. Stimulus novelty, task relevance, and visual evoked-potential in man. Electroen. Clin. Neuro. 39, 131–143 (1975).
Friedman, D., Cycowicz, Y. M. & Gaeta, H. The novelty P3: an event-related brain potential (ERP) sign of the brain’s evaluation of novelty. Neurosci. Biobehav. Rev. 25, 355–373 (2001).
Knight, R. T. Distributed cortical network for visual attention. J Cogn Neurosci 9, 75–91 (1997).
Astikainen, P., Lillstrang, E. & Ruusuvirta, T. Visual mismatch negativity for changes in orientation–a sensory memory-dependent response. Eur. J. Neurosc. 28, 2319–2324 (2008).
Grafton, S. T., Fadiga, L., Arbib, M. A. & Rizzolatti, G. Premotor cortex activation during observation and naming of familiar tools. Neuroimage 6, 231–236 (1997).
Jax, S. A. & Buxbaum, L. J. Response interference between functional and structural actions linked to the same familiar object. Cognition 115, 350–355 (2010).
Miozzo, M., Pulvermüller, F. & Hauk, O. Early parallel activation of semantics and phonology in picture naming: Evidence from a multiple linear regression MEG study. Cereb. Cortex 25, 3343–3355 (2014).
Watson, C. E. & Buxbaum, L. J. A distributed network critical for selecting among tool-directed actions. Cortex 65, 65–82 (2015).
Freeman, W. J. How brains make up their minds. Columbia University Press (2000).
McClelland, J. L., Mirman, D. & Holt, L. L. Are there interactive processes in speech perception? Trends Cogn. Sci. 10, 363–369 (2006).
Cox, D., Meyers, E. & Sinha, P. Contextually evoked object-specific responses in human visual cortex. Science 304, 115–117 (2004).
Baus, C. & Costa, A. On the temporal dynamics of sign production: An ERP study in Catalan Sign Language (LSC). Brain Res. 1609, 40–53 (2015).
Tarr, M. J. & Pinker, S. Mental rotation and orientation-dependence in shape recognition. Cognitive Psychol. 21, 233–282 (1989).
Dingemanse, M., Blasi, D. E., Lupyan, G., Christiansen, M. H. & Monaghan, P. Arbitrariness, iconicity, and systematicity in language. Trends Cogn. Sciences 19, 603–615 (2015).
Lupyan, G. & Thompson-Schill, S. L. The evocative power of words: Activation of concepts by verbal and nonverbal means. J. Exp. Psychol. Gen. 141, 170–186 (2012).
Thompson, R. L., Vinson, D. P., Woll, B. & Vigliocco, G. The road to language learning is iconic: Evidence from British Sign Language. Psychol. Sci. 23, 1443–1448 (2012).
Faul, F., Erdfelder, E., Lang, A. G. & Buchner, A. G* Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav. Res. Methods 39, 175–191 (2007).
Van Selst, M. & Jolicoeur, P. A solution to the effect of sample size on outlier elimination. Q. J. Exp. Psychol.-A 47, 631–650 (1994).
Blythe, H. I., Dickins, J. H., Kennedy, C. R. & Liversedge, S. P. Phonological processing during silent reading in teenagers who are deaf/hard of hearing: An eye movement investigation. Developmental Sci. 21, e12643 (2018).
Kyle, F. E., Campbell, R. & MacSweeney, M. The relative contributions of speechreading and vocabulary to deaf and hearing children’s reading ability. Res. Dev. Disabil. 48, 13–24 (2016).
Faust, M. E., Balota, D. A., Spieler, D. H. & Ferraro, F. R. Individual differences in information-processing rate and amount: implications for group differences in response latency. Psychol. Bull. 125, 777 (1999).
Bates, D., Maechler, M., Bolker, B. & Walker, S. Fitting linear mixed-effects models using lme4. J. Stat. Softw. 67, 1–48 (2015).
R Core Team R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria, https://www.R-project.org/ (2016).
Azizian, A., Freitas, A. L., Watson, T. D. & Squires, N. K. Electrophysiological correlates of categorization: P300 amplitude as index of target similarity. Biol. Psychol. 71, 278–288 (2006).
Delorme, A. & Makeig, S. EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J.Neurosci. Meth. 134, 9–21 (2004).
Stone, J. V. Independent component analysis: an introduction. Trends Cogn. Sci 6, 59–64 (2002).
Tadel, F., Baillet, S., Mosher, J. C., Pantazis, D. & Leahy, R. M. Brainstorm: a user-friendly application for MEG/EEG analysis. Comput. Intel. Neurosc. 2011, 8 (2011).
Oostenveld, R., Fries, P., Maris, E. & Schoffelen, J. M. FieldTrip: open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput. Intel. Neurosc. 2011, 1 (2011).
Katayama, J. I. & Polich, J. Stimulus context determines P3a and P3b. Psychophysiology 35, 23–33 (1998).
Acknowledgements
This research was funded in part by the PRIN-2017 program grant (Ministero dell’Istruzione dell’Università e della Ricerca Prot. 20177894ZH). We thank Elena Pretato for her help in stimulus preparation and Arianna Caccaro for her support in the recruitment of deaf participants.
Author information
Authors and Affiliations
Contributions
All authors developed the study concept and design. Testing, data collection and data analysis were performed by E.N. All authors contributed to data interpretation. M.M. drafted the first version of the manuscript and E.N. and F.P. provided critical revisions and integrations. All authors approved the final version of the manuscript for submission.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Navarrete, E., Miozzo, M. & Peressotti, F. Language can shape the perception of oriented objects. Sci Rep 10, 8409 (2020). https://doi.org/10.1038/s41598-020-65455-6
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-020-65455-6
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.