ASHA journals
Browse
- No file added yet -

The Korean speech recognition sentences (Song et al., 2023)

Download (88.51 kB)
dataset
posted on 2023-09-06, 21:31 authored by Jieun Song, Byunjun Kim, Minjeong Kim, Paul Iverson

Purpose: The aim of this study was to develop and validate a large Korean sentence set with varying degrees of semantic predictability that can be used for testing speech recognition and lexical processing.

Method: Sentences differing in the degree of final-word predictability (predictable, neutral, and anomalous) were created with words selected to be suitable for both native and nonnative speakers of Korean. Semantic predictability was evaluated through a series of cloze tests in which native (n = 56) and nonnative (n = 19) speakers of Korean participated. This study also used a computer language model to evaluate final-word predictabilities; this is a novel approach that the current study adopted to reduce human effort in validating a large number of sentences, which produced results comparable to those of the cloze tests. In a speech recognition task, the sentences were presented to native (n = 23) and nonnative (n = 21) speakers of Korean in speech-shaped noise at two levels of noise.

Results: The results of the speech-in-noise experiment demonstrated that the intelligibility of the sentences was similar to that of related English corpora. That is, intelligibility was significantly different depending on the semantic condition, and the sentences had the right degree of difficulty for assessing intelligibility differences depending on noise levels and language experience. Conclusions: This corpus (1,021 sentences in total) adds to the target languages available in speech research and will allow researchers to investigate a range of issues in speech perception in Korean.

Supplemental Material S1. Full list of sentences.

Song, J., Kim, B., Kim, M., & Iverson, P. (2023). The Korean speech recognition sentences: A large corpus for evaluating semantic context and language experience in speech perception. Journal of Speech, Language, and Hearing Research, 66(9), 3399–3412. https://doi.org/10.1044/2023_JSLHR-23-00137

Funding

This study was supported by the Economic and Social Research Council of the United Kingdom (recipients: Jieun Song and Paul Iverson), the National Research Foundation of Korea grant funded by the Korean Government (Grant 2020S1A5B5A17089211; recipient: Jieun Song), and a grant of the KAIST-KT joint research project through AI2XL Laboratory, Institute of convergence Technology, funded by KT (G01210698, G01220612; recipients: Jieun Song and Minjeong Kim).

History

Usage metrics

    Journal of Speech, Language, and Hearing Research

    Licence

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC