1,803
Views
3
CrossRef citations to date
0
Altmetric
Editorial

Development of Early Multisensory Perception and Communication: From Environmental and Behavioral to Neural Signatures

Multidisciplinary research across the domains of cognitive development, speech, hearing, and educational sciences, field studies in cultural anthropology, and more recently, empirical research in artificial intelligence, computational modeling, and robotics have firmly established that caregiver-infant communication is a multisensory activity. Infants learn about communication and learn to communicate during ongoing interaction with caregivers in a multisensory world. These related domains of investigation, spanning four decades, have provided rich insights into the multisensory properties of caregivers’ facial and vocal expressions, various speech modulations, head and body movements and touch during discourse while naming objects and actions, all embedded in joint attention activities engaging infants to enhance their communicative abilities (Brand, Baldwin, & Ashburn, Citation2002; Gogate, Bahrick, & Watson, Citation2000; Gogate, Bolzani, & Betancourt, Citation2006; Gogate, Maganti, & Bahrick, Citation2015; Harris, Jones, & Grant, Citation1983; Messer, Citation1978; Rohlfing, Fritsch, Wrede, & Jungmann, Citation2006; Smith & Strader, Citation2014; Sullivan & Horowitz, Citation1983; Yu & Smith, Citation2012; Zukow-Goldring, Citation1990). At the same time, experimental (behavioral) research in developmental psychology has established that young infants prefer this multisensory (e.g., auditory-visual) communication over its unisensory (e.g., auditory or visual) components (e.g., Bahrick & Pickens, Citation1988; Patterson & Werker, Citation2003). These experimental studies have underscored that typically developing infants prefer temporally synchronous and spatially collocated faces and voices to asynchronous faces and voices during speech (e.g., Dodd, Citation1979; Hollich, Newman, & Juscyk, Citation2005; Kuhl & Meltzoff, Citation1984; Lewkowicz, Citation2002, Citation2010; Walker, Citation1982), as well as synchronous spoken words and moving referents to asynchronous ones (Gogate, Citation2010; Jesse & Johnson, Citation2012; Rader & Zukow-Goldring, Citation2015).

The neurophysiological correlates of multisensory (auditory-visual) caregiver-infant communication, however, have remained largely unexplored barring a few exceptions. For instance, only recently have researchers started to explore the neurophysiological components of multisensory face-voice perception in infancy (Grossman, Citation2010; Hyde, Jones, Flom, & Porter, Citation2011; Reynolds, Bahrick, Lickliter, & Guy, Citation2014), although earlier research suggested that intersensory face-voice processing was relegated to the left hemisphere (MacKain, Studdert-Kennedy, Spieker, & Stern, Citation1983). Furthermore, although some pioneering neurophysiological research has examined the beginnings of auditory-visual word-object mapping in infants (Asano et al., Citation2015; Friederich & Friederici, Citation2011, Citation2015; Junge, Cutler, & Hagoort, Citation2012; Molfese, Citation1989; Molfese, Morse, & Peters, Citation1990), no studies to date have examined the cortical regions involved in infants’ processing of specific multisensory properties, such as temporal synchrony processing during word-referent mapping when caregivers name objects and events, or infants’ processing of other levels of multisensory discourse. The neural signatures of multisensory communication across key participants, its correlates in caregivers and in infants during ongoing interaction, have received very little attention in empirical research. Although the methods for examining the neural signatures of multisensory communication have been available for decades, researchers have only recently acknowledged the importance of examining the development of these signatures in infants and their caregivers (Tanaka, Fukushima, Okanayo, & Myowa-Yamakoshi, Citation2014), as a means for understanding communicative development.

The present issue aims to highlight this largely unchartered territory of research to better understand the developmental origins of human multisensory communication. Taken together, the two review and four empirical articles emphasize the range and complexity of multisensory mechanisms underlying infant-caregiver dyadic and triadic exchanges in the real world, and multiple facets of their developmental and neurophysiological characteristics. Furthermore, recent theoretical perspectives on brain network development emphasize the need for elucidating the nature of the environment to better comprehend neurophysiological development (e.g., Byrge, Sporns, & Smith, Citation2014), because the brain, body and environment interact with one another on an ongoing basis, influence each other and are continuously being influenced by one another (e.g., James & Maouene, Citation2009; Maouene, Hidaka, & Smith, Citation2008). Thus, the set of papers included in this special issue examine neurophysiological, behavioral and environmental factors to better understand the nature of multisensory communication as a whole.

The present issue begins with a review by Hyde, Flom, and Porter of the neurophysiological and behavioral research to illustrate that infants’ face-voice processing during auditory-visual communication involves general principles of perceptual processing. Their main contention is that despite human communication being a multisensory experience, research and theorizing on infant perception, cognition, and learning related to face-voice perception has largely focused on unimodal or single sense modalities. The same contention holds for research on the neurophysiological underpinnings of auditory-visual communication in general.

Next, Gogate and Hollich review the neurophysiological and behavioral research on noun-object and verb-action mapping across auditory-visual sensory modalities in the early years. Once again, they emphasize the domain-general multisensory perceptual origins and neural underpinnings that give way to language-specific biases in verb and noun learning, eventually leading to children becoming more flexible with noun and verb learning that is less constrained by a noun- or a verb-bias.

The set of empirical papers begins with Borgström, Torkildsen and Lindgren who illustrate how visual perception of familiar objects facilitates toddlers’ neurophysiological responses (N400 amplitude) to pseudo word-referent incongruity. These findings speak to the dependency of language learning on domain-general visual shape perception and simultaneously attest to the multisensory nature of word learning. Next, Suanda, Smith and Yu examine the multisensory properties of caregivers’ extended discourse to toddlers during ongoing parent-child interactions, specifically illustrating several types of redundancies between caregivers’ gestures and their extended conversations as a means for highlighting meaning to toddlers. Complementing Suanda et al.’s empirical findings, Chang, de Barbaro, and Deák examine longitudinally the contingencies between infants’ behaviors with objects and maternal naming. They illustrate increasing redundancies between infants’ gestures, eye-gaze and maternal speech as infants’ develop from 4 to 9 months of age. Chang et al. demonstrate that with increasing infant age, caregivers name objects more often when infants vocalize, or manipulate objects, or gaze at them. The latter two empirical studies underscore the rich contexts that constitute ongoing caregiver-child multisensory communicative interactions, and highlight the ongoing brain-body-environment relations that characterize these interactions. Finally, Patten, Labban, Casenhiser, and Cotton examine the influence of one aspect of this rich context, the absence versus the presence of temporally synchronous faces during naming, to determine their influence on the perception of these linguistic stimuli by children on the autistic spectrum. This research speaks directly to the clinical relevance of multisensory perceptual mechanisms in translational research. Together, these articles underscore the domain-general multisensory nature of human communication in the early years and suggest further research to elucidate the embodied neural signatures of multisensory communication between infants, children, and their caregivers.

Author's Note

LG expresses her deepest appreciation to the authors for their invaluable contributions to this issue, and to the numerous reviewers who generously offered their time and effort to provide expert reviews.

References

  • Asano, M., Imai, M., Kita, S., Kitajo, K., Okada, H., & Thierry, G. (2015). Sound symbolism scaffolds language development in preverbal infants. Cortex, 63(2015), 196–205.
  • Bahrick, L. E., & Pickens, J. N. (1988). Classification of bimodal English and Spanish language passages by infants. Infant Behavior and Development, 11, 277–296.
  • Brand, R., Baldwin, D. A., & Ashburn, L. A. (2002). Evidence for ‘motionese’: Modifications in mothers’ infant-directed action. Developmental Science, 5, 72–83. doi:10.1111/1467-7687.00211
  • Byrge, L., Sporns, O., & Smith, L. B. (2014). Developmental process emerges from extended brain-body-behavior networks. Trends in Cognitive Sciences, 18, 395–403. doi:10.1016/j.tics.2014.04.010
  • Dodd, B. (1979). Lip reading in infants: Attention to speech presented in-and-out of synchrony. Cognitive Psychology, 11, 478–484.
  • Friederich, M., & Friederici, A. D. (2011). Word learning in 6-month-olds: Fast encoding–weak retention. Journal of Cognitive Neuroscience, 23(11), 3228–3240. doi:10.1162/jocn_a_00002
  • Friedrich, M., & Friederici, A. D. (2015). The origins of word learning: Brain responses of three-month-olds indicate their rapid association of objects and words. Developmental Science. doi:10.1111/desc.12357
  • Gogate, L. J. (2010). Learning of syllable-object relations by preverbal infants: The role of temporal synchrony and syllable distinctiveness. Journal of Experimental Child Psychology, 105, 178–197.
  • Gogate, L. J., Bahrick, L. E., & Watson, J. D. (2000). A study of multimodal motherese: The role of temporal synchrony between verbal labels and gestures. Child Development, 71, 878–894. doi:10.1111/cdev.2000.71.issue-4
  • Gogate, L. J., Bolzani, L. H., & Betancourt, E. (2006). Attention to maternal multimodal naming by 6- to 8-month-old infants and learning of word-object relations. Infancy, 9, 259–288. doi:10.1207/s15327078in0903_1
  • Gogate, L. J., Maganti, M., & Bahrick, L. (2015). Cross-cultural evidence for multimodal motherese: Asian-Indian mothers’ adaptive use of synchronous words and gestures. Journal of Experimental Child Psychology, 129, 110–126. doi:10.1016/j.jecp.2014.09.002
  • Grossman, T. (2010). The development of emotion perception in face and voice during infancy. Special issue, Development and plasticity of multisensory functions. Restorative Neurology and Neuroscience, 28(2), 219–236.
  • Harris, M., Jones, D., & Grant, J. (1983). The nonverbal context of mothers’ speech to infants. First Language, 4, 21–30. doi:10.1177/014272378300401003
  • Hollich, G., Newman, R. W., & Juscyk, P. S. (2005). Infants’ use synchronized visual information to separate streams of speech. Child Development, 76(3), 598–613.
  • Hyde, D. C., Jones, B. L., Flom, R., & Porter, C. L. (2011). Neural signatures of face-voice synchrony in 5-month-old human infants. Developmental Psychobiology, 53(4), 359–370. [PubMed: 21271561]. doi:10.1002/dev.v53.4
  • James, K. H., & Maouene, J. (2009). Auditory verb perception recruits motor systems in the developing brain: An fMRI investigation. Developmental Science, 12, F26–F34. doi:10.1111/j.1467-7687.2009.00919.x
  • Jesse, A., & Johnson, E. (2012). Prosodic temporal alignment of co-speech gestures to speech facilitates referent resolution. Journal of Experimental Psychology: Human Perception and Performance, 38(6), 1567–1581.
  • Junge, C., Cutler, A., & Hagoort, P. (2012). Electrophysiological evidence of early word learning. Neuropsychologia, 50, 3702–3712.
  • Kuhl, P. K., & Meltzoff, A. N. (1984). The intermodal representation of speech in infants. Infant Behavior and Development, 7, 361–381.
  • Lewkowicz, D. (2002). Heterogeneity and heterochrony in the development of intersensory perception. Cognitive Brain Research, 14, 41–63. doi:10.1016/S0926-6410(02)00060-5
  • Lewkowicz, D. (2010). Infant perception of audio-visual speech synchrony. Developmental Psychology, 46(1), 66–77.
  • MacKain, K., Studdert-Kennedy, M., Spieker, S., & Stern, D. (1983). Infant intermodal speech perception is a left-hemisphere function. Science, 219(4590), 1347–1349.
  • Maouene, J., Hidaka, S., & Smith, B. L. (2008). Body parts and early-learned verbs. Cognitive Science, 32, 1200–1216. doi:10.1080/03640210802019997
  • Messer, D. J. (1978). The integration of mothers’ referential speech with joint play. Child Development, 49, 781–787. doi:10.2307/1128248
  • Molfese, D. L. (1989). Electrophysiological correlates of word meanings in 14-month-old human infants. Developmental Neuropsychology, 5, 79–103.
  • Molfese, D. L., Morse, P. A., & Peters, C. J. (1990). Auditory evoked responses to names for different objects: Crossmodal processing as a basis for infant language acquisition. Developmental Psychology, 26(5), 780–795.
  • Patterson, M., & Werker, J. F. (2003). Two-month-old infants match phonetic information in lips and voice. Developmental Science, 6, 191–196. doi:10.1111/desc.2003.6.issue-2
  • Rader, N. D., & Zukow-Goldring, P. (2015). The role of speech-gesture synchrony in clipping words from the speech stream: Evidence from infant pupil responses. Ecological Psychology, 27, 290–299. doi:10.1080/10407413.2015.1086226
  • Reynolds, G. D., Bahrick, L. E., Lickliter, R., & Guy, M. W. (2014). Neural correlates of intersensory processing in five-month-old infants. Developmental Psychobiology, 56, 355–372. doi:10.1002/dev.21104
  • Rohlfing, K. J., Fritsch, J., Wrede, B., & Jungmann, T. (2006). How can multimodal cues from child-directed interaction reduce learning complexity in robots? Advanced Robotics, 20, 1183–1199. doi:10.1163/605156855306778522532
  • Smith, N. A., & Strader, H. L. (2014). Infant-directed visual prosody: Mothers’ head movements and speech acoustics. Interaction Studies, 15, 38–54.
  • Sullivan, J. & Horowitz, F. D. (1983). Infant intermodal perception and maternal multimodal stimulation: Implications for language development. In C. Rovee-Collier & L.P. Lipsitt (Eds.), Advances in infancy research, II, (pp. 183–239. Norwood, NJ: Ablex.
  • Tanaka, Y., Fukushima, H., Okanoya, K., & Myowa-Yamakoshi, M. (2014, October). Mothers’ multimodal information processing is modulated by multimodal interactions with their infants. Scientific Reports, 4, 6623. doi:10.1038/srep06623.
  • Walker, A. S. (1982). Intermodal perception of expressive behaviors by human infants. Journal of Experimental Child Psychology, 13, 514–535.
  • Yu, C., & Smith, L. B. (2012). Embodied attention and word learning by toddlers. Cognition, 125, 244–262. doi:10.1016/j.cognition.2012.06.016
  • Zukow-Goldring, P. (1990). Socio-perceptual bases for the emergence of language: An alternative to innatist approaches. Developmental Psychobiology, 23, 705–726.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.