The I2R’s ASR System for the VOiCES from a Distance Challenge 2019

Page view(s)
41
Checked on Mar 25, 2024
The I2R’s ASR System for the VOiCES from a Distance Challenge 2019
Title:
The I2R’s ASR System for the VOiCES from a Distance Challenge 2019
Journal Title:
INTERSPEECH 2019
Publication Date:
15 September 2019
Citation:
Chong, T.Y., Tan, K.M., Teh, K.K., You, C.H., Sun, H., Tran, H.D. (2019) The I2R’s ASR System for the VOiCES from a Distance Challenge 2019. Proc. Interspeech 2019, 2458-2462, DOI: 10.21437/Interspeech.2019-2130.
Abstract:
This paper describes the development of the automatic speech recognition (ASR) system for the submission to the VOiCES from a Distance Challenge 2019. In this challenge, we focused on the fixed condition, where the task is to recognize reverberant and noisy speech based on a limited amount of clean training data. In our system, the mismatch between the training and testing conditions was reduced by using multi-style training where the training data was artificially contaminated with different reverberation and noise sources. Also, the Weighted Prediction Error (WPE) algorithm was used to reduce the reverberant effect in the evaluation data. To boost the system performance, acoustic models of different neural network architectures were trained and the respective systems were fused to give the final output. Moreover, an LSTM language model was used to rescore the lattice to compensate the weak n-gram model trained from only the transcription text. Evaluated on the development set, our system showed an average word error rate (WER) of 27.04%.
License type:
PublisherCopyrights
Funding Info:
Description:
ISBN:

Files uploaded:
File Size Format Action
There are no attached files.