Investigating the Effect of Listening Conditions on Speech Intelligibility

A special issue of Languages (ISSN 2226-471X).

Deadline for manuscript submissions: closed (20 April 2022) | Viewed by 6238

Special Issue Editors


E-Mail Website
Guest Editor
Phonetics Lab, Linguistics Department, UC Davis, Davis, CA 95616, USA
Interests: speech perception; speech production; coarticulation; phonetics and phonology
Department of Linguistics, University of Wisconsin, Milwaukee, WI 53201, USA
Interests: speech perception; speech production; interplay between phonetics; phonology; morphology

Special Issue Information

Dear Colleagues,

The goal of the Special Issue is to bring together work investigating the effect of listening conditions on speech intelligibility. Difficulty in perceiving speech in the presence of background noise has been well attested in the literature, particularly for adults and children with hearing impairment (e.g., Caldwell & Nittrouer, 2013; Wong et al., 2009; Bradlow et al., 2003; Ferguson & Kewley-Port, 2002; Armstrong et al., 1997). Competing auditory signals (e.g., noise, other talkers, etc.) can interfere with a listener’s ability to hear a speaker’s intended message (e.g., Van Engen & Bradlow, 2007). Other work has shown that speakers do not monotonically hyperarticulate all aspects of their speech when speaking clearly, but rather target and maintain helpful contrasts for listeners in ways that can be interpreted as communicatively strategic (Scarborough & Zellou, 2013; Cohn et al., 2021). There is also a great deal of individual variation in speech-in-noise perception (e.g., Bent et al., 2016).

Speech intelligibility is both a practical and a theoretical issue (Smiljanić & Bradlow, 2005). For one, work understanding how speech is produced and perceived in different listening conditions, and to different listener populations, can have practical application (e.g., Adlard & Hazan, 1998), e.g., in clinical settings. In addition, understanding such phenomena is important for theories of cognition because human speech is a remarkably durable system of communication that, despite the wide range of environmental conditions present in everyday life, generally succeeds. Yet, pinpointing exactly how and why it manages to succeed — particularly when confronted with a relatively novel barrier to communication, such as face masks — remains an ongoing challenge for language researchers. The purpose of the Special Issue is to bring together work on this issue.

Examples of research topics suitable for this Special Issue include, but are not limited to:

  • Examination and/or comparisons of the effect of different listening contexts (noise, multitalker-babble, adverse listening conditions, music, etc.) on speech production and/or perception;
  • Explorations of effect of situational context (speaking style, wearing a face mask, complex tasks, attentional factors) on speech-in-noise perception;
  • Investigations of speech intelligibility across different listener groups (children, older adults, those with hearing loss, non-native speakers, etc.);
  • Presentation of novel methods or analyses for examining speech intelligibility in different types of listening conditions;
  • Analysis of speech intelligibility in languages other than English (we encourage submissions focusing on non-Indo European and/or endangered languages);
  • Examination of speech in non-prestige dialects, second languages, or multilingual contexts.

We request that, prior to submitting a manuscript, interested authors initially submit a proposed title and an abstract of 400–600 words summarizing their intended contribution. Please send it to the guest editors (Georgia Zellou, [email protected] and Anne Pycha, [email protected]) or to the Languages editorial office ([email protected]). Abstracts will be reviewed by the guest editors for the purposes of ensuring proper fit within the scope of the Special Issue. Full manuscripts will undergo double-blind peer-review.

Tentative completion schedule:

  • Abstract submission deadline: 01 November 2021
  • Notification of abstract acceptance: 01 January 2022
  • Full manuscript deadline: 20 April 2022

References

Adlard, A., & Hazan, V. (1998). Speech perception in children with specific reading difficulties (dyslexia). The Quarterly Journal of Experimental Psychology Section A51(1), 153-177.

Armstrong, M., Pegg, P., James, C., & Blarney, P. (1997). Speech Perception in Noise With Implant and Hearing Aid. Otology & Neurotology18(6), S140-S141.

Bent, T., Baese-Berk, M., Borrie, S. A., & McKee, M. (2016). Individual differences in the perception of regional, nonnative, and disordered speech varieties. The Journal of the Acoustical Society of America140(5), 3775-3786.

Bradlow, A. R., Kraus, N., & Hayes, E. (2003). Speaking Clearly for Children with Learning Disabilities: Sentence Perception in Noise. Journal of Speech, Language, and Hearing Research46(1), 80-97.

Caldwell, A., & Nittrouer, S. (2012). Speech perception in noise by children with cochlear implants. Journal of Speech, Language, and Hearing Research: JSLHR56(1), 13-30.

Cohn, M., Pycha, A., & Zellou, G. (2021). Intelligibility of face-masked speech depends on speaking style: Comparing casual, clear, and emotional speech. Cognition210, 104570.

Ferguson, S. H., & Kewley-Port, D. (2002). Vowel intelligibility in clear and conversational speech for normal-hearing and hearing-impaired listeners. The Journal of the Acoustical Society of America112(1), 259-271.

Scarborough, R., & Zellou, G. (2013). Clarity in communication: “Clear” speech authenticity and lexical neighborhood density effects in speech production and perception. The Journal of the Acoustical Society of America134(5), 3793-3807.

Smiljanić, R., & Bradlow, A. R. (2009). Speaking and hearing clearly: Talker and listener factors in speaking style changes. Language and linguistics compass3(1), 236-264.

Van Engen, K. J., & Bradlow, A. R. (2007). Sentence recognition in native-and foreign-language multi-talker background noise. The Journal of the Acoustical Society of America121(1), 519-526.

Wong, P. C., Jin, J. X., Gunasekera, G. M., Abel, R., Lee, E. R., & Dhar, S. (2009). Aging and cortical mechanisms of speech perception in noise. Neuropsychologia47(3), 693-703.

Dr. Georgia Zellou
Dr. Anne Pycha
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a double-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Languages is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • speech intelligibility
  • clear speech
  • speech-in-noise perception
  • models of speech production and perception

Published Papers (3 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

14 pages, 1278 KiB  
Article
German Word-Final Devoicing in Naturally-Produced and TTS Speech
by Aleese Block, Kristin Predeck and Georgia Zellou
Languages 2022, 7(4), 270; https://doi.org/10.3390/languages7040270 - 24 Oct 2022
Cited by 1 | Viewed by 1491
Abstract
This study explores the production and perception of word-final devoicing in German across text-to-speech (from technology used in common voice-AI “smart” speaker devices—specifically, voices from Apple and Amazon) and naturally produced utterances. First, the phonetic realization of word-final devoicing in German across text-to-speech [...] Read more.
This study explores the production and perception of word-final devoicing in German across text-to-speech (from technology used in common voice-AI “smart” speaker devices—specifically, voices from Apple and Amazon) and naturally produced utterances. First, the phonetic realization of word-final devoicing in German across text-to-speech (TTS) and naturally produced word productions was compared. Acoustic analyses reveal that the presence of cues to a word-final voicing contrast varied across speech types. Naturally produced words with phonologically voiced codas contain partial voicing, as well as longer vowels than words with voiceless codas. However, these distinctions are not present in TTS speech. Next, German listeners completed a forced-choice identification task, in which they heard the words and made coda consonant categorizations, in order to examine the intelligibility consequences of the word-final devoicing patterns across speech types. Intended coda identifications are higher for the naturally produced productions than for TTS. Moreover, listeners systematically misidentified voiced codas as voiceless in TTS words. Overall, this study extends previous literature on speech intelligibility at the intersection of speech synthesis and contrast neutralization. TTS voices tend to neutralize salient phonetic cues present in natural speech. Subsequently, listeners are less able to identify phonological distinctions in TTS. We also discuss how investigating which cues are more salient in natural speech can be beneficial in synthetic speech generation to make them more natural and also easier to perceive. Full article
Show Figures

Figure 1

22 pages, 1304 KiB  
Article
Predictive Processing and Inhibitory Control Drive Semantic Enhancements for Non-Dominant Language Word Recognition in Noise
by Melinda Fricke and Megan Zirnstein
Languages 2022, 7(3), 239; https://doi.org/10.3390/languages7030239 - 15 Sep 2022
Viewed by 2085
Abstract
Auditory word recognition in the non-dominant language has been suggested to break down under noisy conditions due, in part, to the difficulty of deriving a benefit from contextually constraining information. However, previous studies examining the effects of sentence constraints on word recognition in [...] Read more.
Auditory word recognition in the non-dominant language has been suggested to break down under noisy conditions due, in part, to the difficulty of deriving a benefit from contextually constraining information. However, previous studies examining the effects of sentence constraints on word recognition in noise have conflated multiple psycholinguistic processes under the umbrella term of “predictability”. The present study improves on these by narrowing its focus specifically on prediction processes, and on whether the possibility of using semantic constraint to predict an upcoming target word improves word recognition in noise for different listener populations and noise conditions. We find that heritage, but not second language, Spanish listeners derive a word recognition-in-noise benefit from predictive processing, and that non-dominant language word recognition benefits more from predictive processing under conditions of energetic, rather than informational, masking. The latter suggests that managing interference from competing speech and generating predictions about an upcoming target word draw on the same cognitive resources. An analysis of individual differences shows that better inhibitory control ability is associated with reduced disruption from competing speech in the more dominant language in particular, revealing a critical role for executive function in simultaneously managing interference and generating expectations for upcoming words. Full article
Show Figures

Figure 1

14 pages, 377 KiB  
Article
The Influence of Social Information on Speech Intelligibility within the Spanish Heritage Community
by Cecelia Staggs, Melissa Baese-Berk and Charlie Nagle
Languages 2022, 7(3), 231; https://doi.org/10.3390/languages7030231 - 7 Sep 2022
Viewed by 1813
Abstract
Previous research in speech perception has shown that perception is influenced by social factors that can result in behavioral consequences such as reduced intelligibility (i.e., a listeners’ ability to transcribe the speech they hear). However, little is known about these effects regarding Spanish [...] Read more.
Previous research in speech perception has shown that perception is influenced by social factors that can result in behavioral consequences such as reduced intelligibility (i.e., a listeners’ ability to transcribe the speech they hear). However, little is known about these effects regarding Spanish speakers’ perception of heritage Spanish, Spanish spoken by individuals who have an ancestral and cultural connection to the Spanish language. Given that ideologies within the U.S. Latino community often equate Latino identity to speaking Spanish “correctly” and proficiently, there is a clear need to understand the potential influence these ideologies have on speech perception. Using a matched-guised methodology, we analyzed the influence of speaker social background information and listener social background information on speech perception. Participants completed a transcription task in which four different Spanish heritage speakers were paired with different social guises to determine if the speakers were perceived as equally intelligible under each guise condition. The results showed that social guise and listener social variables did not significantly predict intelligibility scores. We argue that the unique socio-political culture within the U.S. Latino community may lead to different effects of language ideology and social expectation on speech perception than what has been documented in previous work. Full article
Back to TopTop