Abstract

Existing speech classification algorithms often perform well when evaluated on training and test data drawn from the same distribution. In practice, however, these distributions are not always the same. In these circumstances, the performance of trained models will likely decrease. In this paper, we discuss an underutilized divergence measure and derive an estimable upper bound on the test error rate that depends on the error rate on the training data and the distance between training and test distributions. Using this bound as motivation, we develop a feature learning algorithm that aims to identify invariant speech features that generalize well to data similar to, but different from, the training set. Comparative results confirm the efficacy of the algorithm on a set of cross-domain speech classification tasks.

Original languageEnglish (US)
Title of host publication2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages77-82
Number of pages6
ISBN (Print)9781479971299
DOIs
StatePublished - Apr 1 2014
Event2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - South Lake Tahoe, United States
Duration: Dec 7 2014Dec 10 2014

Other

Other2014 IEEE Workshop on Spoken Language Technology, SLT 2014
CountryUnited States
CitySouth Lake Tahoe
Period12/7/1412/10/14

Fingerprint

Learning algorithms
Divergence
Efficacy

Keywords

  • Domain adaptation
  • Feature selection
  • Machine learning
  • Pathological speech analysis

ASJC Scopus subject areas

  • Computer Science Applications
  • Human-Computer Interaction
  • Computer Vision and Pattern Recognition
  • Artificial Intelligence
  • Language and Linguistics

Cite this

Wisler, A., Berisha, V., Liss, J., & Spanias, A. (2014). Domain invariant speech features using a new divergence measure. In 2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings (pp. 77-82). [7078553] Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.1109/SLT.2014.7078553

Domain invariant speech features using a new divergence measure. / Wisler, Alan; Berisha, Visar; Liss, Julie; Spanias, Andreas.

2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2014. p. 77-82 7078553.

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Wisler, A, Berisha, V, Liss, J & Spanias, A 2014, Domain invariant speech features using a new divergence measure. in 2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings., 7078553, Institute of Electrical and Electronics Engineers Inc., pp. 77-82, 2014 IEEE Workshop on Spoken Language Technology, SLT 2014, South Lake Tahoe, United States, 12/7/14. https://doi.org/10.1109/SLT.2014.7078553
Wisler A, Berisha V, Liss J, Spanias A. Domain invariant speech features using a new divergence measure. In 2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings. Institute of Electrical and Electronics Engineers Inc. 2014. p. 77-82. 7078553 https://doi.org/10.1109/SLT.2014.7078553
Wisler, Alan ; Berisha, Visar ; Liss, Julie ; Spanias, Andreas. / Domain invariant speech features using a new divergence measure. 2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2014. pp. 77-82
@inproceedings{f2c8c07e0ec847eb9f30b3ce8d14d9ba,
title = "Domain invariant speech features using a new divergence measure",
abstract = "Existing speech classification algorithms often perform well when evaluated on training and test data drawn from the same distribution. In practice, however, these distributions are not always the same. In these circumstances, the performance of trained models will likely decrease. In this paper, we discuss an underutilized divergence measure and derive an estimable upper bound on the test error rate that depends on the error rate on the training data and the distance between training and test distributions. Using this bound as motivation, we develop a feature learning algorithm that aims to identify invariant speech features that generalize well to data similar to, but different from, the training set. Comparative results confirm the efficacy of the algorithm on a set of cross-domain speech classification tasks.",
keywords = "Domain adaptation, Feature selection, Machine learning, Pathological speech analysis",
author = "Alan Wisler and Visar Berisha and Julie Liss and Andreas Spanias",
year = "2014",
month = "4",
day = "1",
doi = "10.1109/SLT.2014.7078553",
language = "English (US)",
isbn = "9781479971299",
pages = "77--82",
booktitle = "2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings",
publisher = "Institute of Electrical and Electronics Engineers Inc.",

}

TY - GEN

T1 - Domain invariant speech features using a new divergence measure

AU - Wisler, Alan

AU - Berisha, Visar

AU - Liss, Julie

AU - Spanias, Andreas

PY - 2014/4/1

Y1 - 2014/4/1

N2 - Existing speech classification algorithms often perform well when evaluated on training and test data drawn from the same distribution. In practice, however, these distributions are not always the same. In these circumstances, the performance of trained models will likely decrease. In this paper, we discuss an underutilized divergence measure and derive an estimable upper bound on the test error rate that depends on the error rate on the training data and the distance between training and test distributions. Using this bound as motivation, we develop a feature learning algorithm that aims to identify invariant speech features that generalize well to data similar to, but different from, the training set. Comparative results confirm the efficacy of the algorithm on a set of cross-domain speech classification tasks.

AB - Existing speech classification algorithms often perform well when evaluated on training and test data drawn from the same distribution. In practice, however, these distributions are not always the same. In these circumstances, the performance of trained models will likely decrease. In this paper, we discuss an underutilized divergence measure and derive an estimable upper bound on the test error rate that depends on the error rate on the training data and the distance between training and test distributions. Using this bound as motivation, we develop a feature learning algorithm that aims to identify invariant speech features that generalize well to data similar to, but different from, the training set. Comparative results confirm the efficacy of the algorithm on a set of cross-domain speech classification tasks.

KW - Domain adaptation

KW - Feature selection

KW - Machine learning

KW - Pathological speech analysis

UR - http://www.scopus.com/inward/record.url?scp=84946691492&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84946691492&partnerID=8YFLogxK

U2 - 10.1109/SLT.2014.7078553

DO - 10.1109/SLT.2014.7078553

M3 - Conference contribution

AN - SCOPUS:84946691492

SN - 9781479971299

SP - 77

EP - 82

BT - 2014 IEEE Workshop on Spoken Language Technology, SLT 2014 - Proceedings

PB - Institute of Electrical and Electronics Engineers Inc.

ER -