Publication Details
Neural Network Bottleneck Features for Language Identification
Zhang Le
Ng Tim
Mallidi Sri Harish
Glembek Ondřej, Ing., Ph.D.
Ma Jeff
Zhang Bing
language identification, noisy speech, robustfeature extraction
We have presented the bottleneck features in the context of Languageidentification. It combines benefits of both phonotacticand acoustic system. Usually, the phonotactic system is favorablefor the long duration files, while acoustic for the short ones.This approach takes the advantage of both. In addition, we canalso use modeling of context dependent phonemes in bottleneckfeatures. This brings very nice improvement over the contextindependent phonemes.
This paper presents the application of Neural Network Bottleneck (BN) features in Language Identification (LID). BN features are generally used for Large Vocabulary Speech Recognition in conjunction with conventional acoustic features, such as MFCC or PLP.We compare the BN features to several common types of acoustic features used in the state-of-the-art LID systems. The test set is from DARPA RATS (Robust Automatic Transcription of Speech) program, which seeks to advance state-of-the-art detection capabilities on audio from highly degraded radio communication channels. On this type of noisy data, we show that in average, the BN features provide a 45% relative improvement in the Cavgor Equal Error Rate (EER) metrics across several test duration conditions, with respect to our single best acoustic features.
@inproceedings{BUT111630,
author="Pavel {Matějka} and Le {Zhang} and Tim {Ng} and Sri Harish {Mallidi} and Ondřej {Glembek} and Jeff {Ma} and Bing {Zhang}",
title="Neural Network Bottleneck Features for Language Identification",
booktitle="Proceedings of Odyssey 2014",
year="2014",
journal="Proceedings of Odyssey: The Speaker and Language Recognition Workshop Odyssey 2014, Joensuu, Finland",
volume="2014",
number="6",
pages="299--304",
publisher="International Speech Communication Association",
address="Joensuu",
issn="2312-2846",
url="http://www.fit.vutbr.cz/research/groups/speech/publi/2014/matejka_odyssey2014_299-304-35.pdf"
}