TY - JOUR
T1 - Objective intelligibility assessment by automated segmental and suprasegmental listening error analysis
AU - Jiao, Yishan
AU - LaCross, Amy
AU - Berisha, Visar
AU - Liss, Julie
N1 - Funding Information:
The authors report funding from the Foundation for the National Institutes of Health under Grant 2 R01 DC006859-11 (awarded to PI: Julie Liss and MPI: Visar Berisha).
Publisher Copyright:
© 2019 American Speech-Language-Hearing Association.
PY - 2019/9
Y1 - 2019/9
N2 - Purpose: Subjective speech intelligibility assessment is often preferred over more objective approaches that rely on transcript scoring. This is, in part, because of the intensive manual labor associated with extracting objective metrics from transcribed speech. In this study, we propose an automated approach for scoring transcripts that provides a holistic and objective representation of intelligibility degradation stemming from both segmental and suprasegmental contributions, and that corresponds with human perception. Method: Phrases produced by 73 speakers with dysarthria were orthographically transcribed by 819 listeners via Mechanical Turk, resulting in 63,840 phrase transcriptions. A protocol was developed to filter the transcripts, which were then automatically analyzed using novel algorithms developed for measuring phoneme and lexical segmentation errors. The results were compared with manual labels on a randomly selected sample set of 40 transcribed phrases to assess validity. A linear regression analysis was conducted to examine how well the automated metrics predict a perceptual rating of severity and word accuracy. Results: On the sample set, the automated metrics achieved 0.90 correlation coefficients with manual labels on measuring phoneme errors, and 100% accuracy on identifying and coding lexical segmentation errors. Linear regression models found that the estimated metrics could predict a significant portion of the variance in perceptual severity and word accuracy. Conclusions: The results show the promising development of an objective speech intelligibility assessment that identifies intelligibility degradation on multiple levels of analysis.
AB - Purpose: Subjective speech intelligibility assessment is often preferred over more objective approaches that rely on transcript scoring. This is, in part, because of the intensive manual labor associated with extracting objective metrics from transcribed speech. In this study, we propose an automated approach for scoring transcripts that provides a holistic and objective representation of intelligibility degradation stemming from both segmental and suprasegmental contributions, and that corresponds with human perception. Method: Phrases produced by 73 speakers with dysarthria were orthographically transcribed by 819 listeners via Mechanical Turk, resulting in 63,840 phrase transcriptions. A protocol was developed to filter the transcripts, which were then automatically analyzed using novel algorithms developed for measuring phoneme and lexical segmentation errors. The results were compared with manual labels on a randomly selected sample set of 40 transcribed phrases to assess validity. A linear regression analysis was conducted to examine how well the automated metrics predict a perceptual rating of severity and word accuracy. Results: On the sample set, the automated metrics achieved 0.90 correlation coefficients with manual labels on measuring phoneme errors, and 100% accuracy on identifying and coding lexical segmentation errors. Linear regression models found that the estimated metrics could predict a significant portion of the variance in perceptual severity and word accuracy. Conclusions: The results show the promising development of an objective speech intelligibility assessment that identifies intelligibility degradation on multiple levels of analysis.
UR - http://www.scopus.com/inward/record.url?scp=85072546655&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85072546655&partnerID=8YFLogxK
U2 - 10.1044/2019_JSLHR-S-19-0119
DO - 10.1044/2019_JSLHR-S-19-0119
M3 - Article
C2 - 31525112
AN - SCOPUS:85072546655
SN - 1092-4388
VL - 62
SP - 3359
EP - 3366
JO - Journal of Speech and Hearing Disorders
JF - Journal of Speech and Hearing Disorders
IS - 9
ER -