跳到主要导航 跳到搜索 跳到主要内容

Applying batch normalization to hybrid NN-HMM model for speech recognition

  • Hongjian Zhan*
  • , Guilin Chen
  • , Yue Lu
  • *此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Batch Normalization has showed success in image classification and other image processing areas by reducing internal covariate shift in deep network model’s training procedure. In this paper, we propose to apply batch normalization to speech recognition within the hybrid NNHMM model. We evaluate the performance of this new method in the acoustic model of the hybrid system with a speaker-independent speech recognition task using some Chinese datasets. Compared to the former best model we used in the Chinese datasets, it shows that with batch normalization we can reach lower word error rate (WER) of 8%–13% relatively, meanwhile we just need 60% iterations of original model to finish the training procedure.

源语言英语
主期刊名Pattern Recognition - 7th Chinese Conference, CCPR 2016, Proceedings
编辑Tieniu Tan, Xilin Chen, Xuelong Li, Jian Yang, Hong Cheng, Jie Zhou
出版商Springer Verlag
427-435
页数9
ISBN(印刷版)9789811030048
DOI
出版状态已出版 - 2016

出版系列

姓名Communications in Computer and Information Science
663
ISSN(印刷版)1865-0929

指纹

探究 'Applying batch normalization to hybrid NN-HMM model for speech recognition' 的科研主题。它们共同构成独一无二的指纹。

引用此