z-logo
open-access-imgOpen Access
Deeper Time Delay Neural Networks for Effective Acoustic Modelling
Author(s) -
Xiaorong Huang,
Weibin Zhang,
Xiangmin Xu,
Rong Yin,
Dongpeng Chen
Publication year - 2019
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/1229/1/012076
Subject(s) - overfitting , dropout (neural networks) , computer science , artificial neural network , salient , time delay neural network , speech recognition , layer (electronics) , artificial intelligence , machine learning , chemistry , organic chemistry
Time delay neural networks (TDNNs) have been shown to be an efficient network architecture for modelling long temporal contexts in speech recognition. Meanwhile, the training times of TDNNs are much less, compared with other long temporal contexts models based on recurrent neural networks. In this paper, we propose deeper architectures to improve the modelling power of TDNNs. At each TDNN layer that needs spliced input, we increase the number of transforms so that the lower layers can provide more salient features for upper layers. Dropout is found to be an effective way to prevent the model from overfitting once the depth of the model is substantially increased. The proposed architectures significantly improvements the recognition accuracy in Switchboard and AMI.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here