314
Views
4
CrossRef citations to date
0
Altmetric
REGULAR ARTICLES

Word segmentation from noise-band vocoded speech

, , &
Pages 1344-1356 | Received 08 Jul 2016, Accepted 02 Jul 2017, Published online: 20 Jul 2017

References

  • Arciuli, J., & Simpson, I. C. (2012). Statistical learning is related to reading ability in children and adults. Cognitive Science, 36, 286–304. doi: 10.1111/j.1551-6709.2011.01200.x
  • Aslin, R. N., Saffran, J. R., & Newport, E. L. (1998). Computation of conditional probability statistics by 8-month-old infants. Psychological Science, 9, 321–324. doi: 10.1111/1467-9280.00063
  • Aslin, R. N., Woodward, J. Z., LaMendola, N. P., & Bever, T. G. (1996). Models of word segmentation in fluent maternal speech to infants. Signal to Syntax: Bootstrapping from Speech to Grammar in Early Acquisition, 117–134.
  • Batterink, L., & Neville, H. (2013). Implicit and explicit second language training recruit common neural mechanisms for syntactic processing. Journal of Cognitive Neuroscience, 25(6), 936–951. doi: 10.1162/jocn_a_00354
  • Boersma, P., & Weenink, D. (2009). Praat: Doing phonetics by computer (Version 5.1. 10)[Computer program]. Retrieved July 8, 2009.
  • Bortfeld, H., Morgan, J. L., Golinkoff, R. M., & Rathbun, K. (2005). Mommy and me familiar names help launch babies into speech-stream segmentation. Psychological Science, 16(4), 298–304. doi: 10.1111/j.0956-7976.2005.01531.x
  • Brent, M. R., & Cartwright, T. A. (1996). Distributional regularity and phonotactic constraints are useful for segmentation. Cognition, 61(1), 93–125. doi: 10.1016/S0010-0277(96)00719-6
  • Brent, M. R., & Siskind, J. M. (2001). The role of exposure to isolated words in early vocabulary development. Cognition, 81(2), B33–B44. doi: 10.1016/S0010-0277(01)00122-6
  • Broadbent, D. E. (1958). Perception and communication. Elmsford, NY: Pergamon Press. doi: 10.1037/10037-000
  • Church, R., Bernhardt, B., Shi, R., & Pichora-Fuller, K. (2005). Infant-directed speech: Final syllable lengthening and rate of speech. The Journal of the Acoustical Society of America, 117(4.2), 2429–2430. doi: 10.1121/1.4786663
  • Cunillera, T., Càmara, E., Laine, M., & Rodríguez-Fornells, A. (2010a). Speech segmentation is facilitated by visual cues. The Quarterly Journal of Experimental Psychology, 63(2), 260–274. doi: 10.1080/17470210902888809
  • Cunillera, T., Càmara, E., Laine, M., & Rodríguez-Fornells, A. (2010b). Words as anchors. Experimental Psychology, 57(2), 134–141. doi: 10.1027/1618-3169/a000017
  • Cunillera, T., Laine, M., & Rodríguez-Fornells, A. (2016). Headstart for speech segmentation: A neural signature for the anchor word effect. Neuropsychologia, 82, 189–199. doi: 10.1016/j.neuropsychologia.2016.01.011
  • Dahan, D., & Brent, M. R. (1999). On the discovery of novel word like units from utterances: An artificial-language study with implications for native-language acquisition. Journal of Experimental Psychology: General, 128(2), 165–185. doi: 10.1037/0096-3445.128.2.165
  • Davis, M. H., & Johnsrude, I. S. (2007). Hearing speech sounds: Top-down influences on the interface between audition and speech perception. Hearing Research, 229(1–2), 132–147. doi: 10.1016/j.heares.2007.01.014
  • Davis, M. H., Johnsrude, I. S., Hervais-Adelman, A. G., Taylor, K., & McGettigan, C. (2005). Lexical information drives perceptual learning of distorted speech: Evidence from the comprehension of noisevocoded sentences. Journal of Experimental Psychology: General, 134(2), 222–241. doi: 10.1037/0096-3445.134.2.222
  • Donaldson, G. S., & Kreft, H. A. (2006). Effects of vowel context on the recognition of initial and medial consonants by cochlear implant users. Ear and Hearing, 27(6), 658–677. doi: 10.1097/01.aud.0000240543.31567.54
  • Downs, D. W., & Crum, M. A. (1978). Processing demands during auditory learning under degraded listening conditions. Journal of Speech, Language, and Hearing Research, 21(4), 702–714. doi: 10.1044/jshr.2104.702
  • Estes, K. G., Evans, J. L., Alibali, M. W., & Saffran, J. R. (2007). Can infants map meaning to newly segmented words? Statistical segmentation and word learning. Psychological Science, 18(3), 254–260. doi: 10.1111/j.1467-9280.2007.01885.x
  • Fiser, J., & Aslin, R. N. (2002). Statistical learning of higher-order temporal structure from visual shape sequences. Journal of Experimental Psychology: Learning, Memory, and Cognition, 28(3), 458–467. doi: 10.1037/0278-7393.28.3.458
  • Frank, M. C., Tenenbaum, J. B., Gibson, E., & Snyder J. (2013). Learning and long-term retention of large-scale artificial languages. PLoS One, 8(1), e52500. doi: 10.1371/journal.pone.0052500
  • Fu, Q.-J., & Nogaki, G. (2005). Noise susceptibility of cochlear implant users: The role of spectral resolution and smearing. Journal of the Association for Research in Otolaryngology, 6(1), 19–27. doi: 10.1007/s10162-004-5024-3
  • Gordon-Salant, S., Yeni-Komshian, G. H., Fitzgibbons, P. J., & Cohen, J. I. (2015). Effects of age and hearing loss on recognition of unaccented and accented multisyllabic words. The Journal of the Acoustical Society of America, 137(2), 884–897. doi: 10.1121/1.4906270
  • Graf Estes, K., & Lew-Williams, C. (2015). Listening through voices: Infant statistical word segmentation across multiple speakers. Developmental Psychology, 51(11), 1517–1528. doi: 10.1037/a0039725
  • Greenwood, D. D. (1990). A cochlear frequency-position function for several species – 29 years later. The Journal of the Acoustical Society of America, 87(6), 2592–2605. doi: 10.1121/1.399052
  • Grieco-Calub, T. M., Saffran, J. R., & Litovsky, R. Y. (2009). Spoken word recognition in toddlers who use cochlear implants. Journal of Speech Language and Hearing Research, 52(6), 1390–1400. doi:0.1044/1092-4388(2009/08-0154) doi: 10.1044/1092-4388(2009/08-0154)
  • Grieco-Calub, T. M., Ward, K. M., & Brehm, L. (2017). Multitasking during degraded speech recognition in school-Age children. Trends Hear, 21, 1–14. doi: 10.1177/2331216516686786
  • Hervais-Adelman, A., Davis, M. H., Johnsrude, I. S., & Carlyon, R. P. (2008). Perceptual learning of noise vocoded words: Effects of feedback and lexicality. Journal of Experimental Psychology: Human Perception and Performance, 34(2), 460–474. doi: 10.1037/0096-1523.34.2.460
  • Hervais-Adelman, A. G., Carlyon, R. P., Johnsrude, I. S., & Davis, M. H. (2012). Brain regions recruited for the effortful comprehension of noise-vocoded words. Language and Cognitive Processes, 27(7-8), 1145–1166. doi: 10.1080/01690965.2012.662280
  • Hervais-Adelman, A. G., Davis, M. H., Johnsrude, I. S., Taylor, K. J., & Carlyon, R. P. (2011). Generalization of perceptual learning of vocoded speech. Journal of Experimental Psychology: Human Perception and Performance, 37(1), 283–295. doi: 10.1037/a0020772
  • Houston, D. M., Jusczyk, P. W., Kuijpers, C., Coolen, R., & Cutler, A. (2000). Cross-language word segmentation by 9-month-olds. Psychonomic Bulletin & Review, 7(3), 504–509. doi: 10.3758/BF03214363
  • Johnson, E. K., & Jusczyk, P. W. (2001). Word segmentation by 8-month-olds: When speech cues count more than statistics. Journal of Memory and Language, 44(4), 548–567. doi: 10.1006/jmla.2000.2755
  • Johnson, E. K., Seidl, A., Tyler, M. D., & Berwick R. C. (2014). The edge factor in early word segmentation: Utterance-level prosody enables word form extraction by 6-month-olds. PloS one, 9(1), e83546. doi: 10.1371/journal.pone.0083546
  • Jusczyk, P. W. (1999). How infants begin to extract words from speech. Trends in Cognitive Sciences, 3(9), 323–328. doi: 10.1016/S1364-6613(99)01363-7
  • Jusczyk, P. W., & Aslin, R. N. (1995). Infants′ detection of the sound patterns of words in fluent speech. Cognitive Psychology, 29(1), 1–23. doi: 10.1006/cogp.1995.1010
  • Kirkham, N. Z., Slemmer, J. A., & Johnson, S. P. (2002). Visual statistical learning in infancy: Evidence for a domain general learning mechanism. Cognition, 83(2), B35–B42. doi: 10.1016/S0010-0277(02)00004-5
  • Lany, J., Shoaib, A., Thompson, A., & Graf Estes, K. (2016). Is statistical-learning ability related to real-time language processing? In Proceedings of the 40th Annual Boston University Conference on Language Development (pp. 203–215). Somerville, MA: Cascadilla Press.
  • Lew-Williams, C., Pelucchi, B., & Saffran, J. R. (2011). Isolated words enhance statistical language learning in infancy. Developmental Science, 14(6), 1323–1329. doi: 10.1111/j.1467-7687.2011.01079.x
  • Lew-Williams, C., & Saffran, J. R. (2012). All words are not created equal: Expectations about word length guide infant statistical learning. Cognition, 122(2), 241–246. doi: 10.1016/j.cognition.2011.10.007
  • Liu, S., & Zheng, F. (2006). Temporal properties in clear speech perception. The Journal of the Acoustical Society of America, 120(1), 424–432. doi: 10.1121/1.2208427
  • Mattys, S. L. (2004). Stress versus coarticulation: Toward an integrated approach to explicit speech segmentation. Journal of Experimental Psychology: Human Perception and Performance, 30(2), 397–408. doi: 10.1037/0096-1523.30.2.397
  • Mattys, S. L., Davis, M. H., Bradlow, A. R., & Scott, S. K. (2012). Speech recognition in adverse conditions: A review. Language and Cognitive Processes, 27(7–8), 953–978. doi: 10.1080/01690965.2012.705006
  • McCoy, S. L., Tun, P. A., Cox, L. C., Colangelo, M., Stewart, R. A., & Wingfield, A. (2005). Hearing loss and perceptual effort: Downstream effects on older adults’ memory for speech. The Quarterly Journal of Experimental Psychology Section A, 58, 22–33. doi: 10.1080/02724980443000151
  • McMillan, B. T. M., & Saffran, J. R. (2016). Learning in complex environments: The effects of background speech on early word learning. Child Development, 87(6), 1841–1855. doi: 10.1111/cdev.12559
  • Mersad, K., & Nazzi, T. (2012). When mommy comes to the rescue of statistics: Infants combine top-down and bottom-up cues to segment speech. Language Learning and Development, 8, 303–315. doi: 10.1080/15475441.2011.609106
  • Misyak, J. B., & Christiansen, M. H. (2012). Statistical learning and language: An individual differences study. Language Learning, 62(1), 302–331. doi: 10.1111/j.1467-9922.2010.00626.x
  • Morgan-Short, K., Steinhauer, K., Sanz, C., & Ullman, M. T. (2012). Explicit and implicit second language training differentially affect the achievement of native-like brain activation patterns. Journal of Cognitive Neuroscience, 24(4), 933–947. doi: 10.1162/jocn_a_00119
  • Munson, B., Donaldson, G. S., Allen, S. L., Collison, E. A., & Nelson, D. A. (2003). Patterns of phoneme perception errors by listeners with cochlear implants as a function of overall speech perception ability. The Journal of the Acoustical Society of America, 113(2), 925–935. doi: 10.1121/1.1536630
  • Norris, J. M., & Ortega, L. (2000). Effectiveness of L2 instruction: A research synthesis and quantitative meta-analysis. Language Learning, 50(3), 417–528. doi: 10.1111/0023-8333.00136
  • Obleser, J., Wise, R. J. S., Dresner, M. A., & Scott, S. K. (2007). Functional integration across brain regions improves speech perception under adverse listening conditions. Journal of Neuroscience, 27(9), 2283–2289. doi: 10.1523/JNEUROSCI.4663-06.2007
  • Palmer, S. D., & Mattys, S. L. (2016). Speech segmentation by statistical learning is supported by domain-general processes within working memory. The Quarterly Journal of Experimental Psychology, 1–12. doi: 10.1080/17470218.2015.1112825
  • Pals, C., Sarampalis, A., & Başkent, D. (2013). Listening effort with cochlear implant simulations. Journal of Speech Language and Hearing Research, 56(4), 1075–1084. doi: 10.1044/1092-4388(2012/12-0074)
  • Pelucchi, B., Hay, J. F., & Saffran, J. R. (2009). Statistical learning in a natural language by 8-month-old infants. Child Development, 80, 674–685. doi: 10.1111/j.1467-8624.2009.01290.x
  • Peña, M., Bonatti, L. L., Nespor, M., & Mehler, J. (2002). Signal-driven computations in speech processing. Science, 298(5593), 604–607. doi: 10.1126/science.1072901
  • Pichora-Fuller, M. K., Schneider, B. A., & Daneman, M. (1995). How young and old adults listen to and remember speech in noise. The Journal of the Acoustical Society of America, 97(1), 593–608. doi: 10.1121/1.412282
  • Rabbitt, P. M. (1966). Recognition: Memory for words correctly heard in noise. Psychonomic Science, 6(8), 383–384. doi: 10.3758/BF03330948
  • Rakerd, B., Seitz, P., & Whearty, M. (1996). Assessing the cognitive demands of speech listening for people with hearing losses. Ear and Hearing, 17(2), 97–106. doi: 10.1097/00003446-199604000-00002
  • R Core Team. (2012). R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. ISBN 3-900051-07-0, URL http://www.R-project.org/.
  • Romberg, A. R., & Saffran, J. R. (2010). Statistical learning and language acquisition. Wiley Interdisciplinary Reviews: Cognitive Science, 1(6), 906–914. doi: 10.1002/wcs.78
  • Rönnberg, J., Lunner, T., Zekveld, A., et al. (2013). The ease of language understanding (ELU) model: Theory, data, and clinical implications. Frontiers in Systems Neuroscience, 7, 1–17. doi: 10.3389/fnsys.2013.00031
  • Saffran, J. R., Aslin, R. N., & Newport, E. L. (1996). Statistical learning by eight-month old infants. Science, 274, 1926–1928. doi: 10.1126/science.274.5294.1926
  • Saffran, J. R., Newport, E. L., & Aslin, R. N. (1996). Word segmentation: The role of distributional cues. Journal of Memory and Language, 35(4), 606–621. doi: 10.1006/jmla.1996.0032
  • Saffran, J. R., & Thiessen, E. D. (2003). Pattern induction by infant language learners. Developmental Psychology, 39(3), 484–494. doi: 10.1037/0012-1649.39.3.484
  • Sarampalis, A., Kalluri, S., Edwards, B., & Hafter, E. (2009). Objective measures of listening effort: Effects of background noise and noise reduction. Journal of Speech Language and Hearing Research, 52(5), 1230–1240. doi: 10.1044/1092-4388(2009/08-0111)
  • Seidl, A., & Johnson, E. K. (2006). Infant word segmentation revisited: Edge alignment facilitates target extraction. Developmental Science, 9(6), 565–573. doi: 10.1111/j.1467-7687.2006.00534.x
  • Shannon, R. V., Zeng, F. G., Kamath, V., Wygonski, J., & Ekelid, M. (1995). Speech recognition with primarily temporal cues. Science, 270(5234), 303–304. doi: 10.1126/science.270.5234.303
  • Shukla, M., Nespor, M., & Mehler, J. (2007). An interaction between prosody and statistics in the segmentation of fluent speech. Cognitive Psychology, 54(1), 1–32. doi: 10.1016/j.cogpsych.2006.04.002
  • Soderstrom, M., Seidl, A., Nelson, D. G. K., & Jusczyk, P. W. (2003). The prosodic bootstrapping of phrases: Evidence from prelinguistic infants. Journal of Memory and Language, 49(2), 249–267. doi: 10.1016/S0749-596X(03)00024-X
  • Sohoglu, E., Peelle, J. E., Carlyon, R. P., & Davis, M. H. (2012). Predictive top-down integration of prior knowledge during speech perception. Journal of Neuroscience, 32(25), 8443–8453. doi: 10.1523/JNEUROSCI.5069-11.2012
  • Spencer, M., Kaschak, M. P., Jones, J. L., & Lonigan, C. J. (2015). Statistical learning is related to early literacy-related skills. Reading and Writing, 28(4), 467–490. doi: 10.1007/s11145-014-9533-0
  • Thiessen, E. D., & Saffran, J. R. (2003). When cues collide: Use of stress and statistical cues to word boundaries by 7-to 9-month-old infants. Developmental Psychology, 39(4), 706–716. doi: 10.1037/0012-1649.39.4.706
  • Toro, J. M., Pons, F., Bion, R. A., & Sebastián-Gallés, N. (2011). The contribution of language-specific knowledge in the selection of statistically-coherent word candidates. Journal of Memory and Language, 64(2), 171–180. doi: 10.1016/j.jml.2010.11.005
  • Ward, K. M., Shen, J., Souza, P. E., & Grieco-Calub, T. M. (2017). Age-related differences in listening effort during degraded speech recognition. Ear and Hearing, 38(1), 74–84. 10.1097/AUD.0000000000000355 doi: 10.1097/AUD.0000000000000355
  • Wild, C. J., Yusuf, A., Wilson, D. E., Peelle, J. E., Davis, M. H., & Johnsrude, I. S. (2012). Effortful listening: The processing of degraded speech depends critically on attention. Journal of Neuroscience, 32(40), 14010–14021. doi: 10.1523/JNEUROSCI.1528-12.2012
  • Winn, M. (2016). Rapid release from listening effort resulting from semantic context, and effects of spectral degradation and cochlear implants. Trends in Hearing, 20, 1–17. doi: 10.1177/2331216516669723
  • Winn, M. B., Edwards, J. R., & Litovsky, R. Y. (2015). The impact of auditory spectral resolution on listening effort revealed by pupil dilation. Ear and Hearing, 36(4), e153-e165. doi: 10.1097/AUD.0000000000000145
  • Xu, L., & Pfingst, B. E. (2008). Spectral and temporal cues for speech recognition: Implications for auditory prostheses. Hearing Research, 242(1), 132–140. doi: 10.1016/j.heares.2007.12.010
  • Zhou, N., Xu, L., & Lee, C.-Y. (2010). The effects of frequency-place shift on consonant confusion in cochlear implant simulations. The Journal of the Acoustical Society of America, 128(1), 401–409. doi: 10.1121/1.3436558

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.