应用科学学报 ›› 2015, Vol. 33 ›› Issue (3): 253-261.doi: 10.3969/j.issn.0255-8297.2015.03.004

• 通信工程 • 上一篇    下一篇

基于循环神经网络的汉语语言模型并行优化算法

王龙1,2, 杨俊安1,2, 陈雷1,2, 林伟3, 刘辉1,2   

  1. 1. 电子工程学院,合肥230037
    2. 安徽省电子制约技术重点实验室,合肥230037
    3. 安徽科大讯飞公司,合肥230037
  • 收稿日期:2014-12-25 修回日期:2015-03-04 出版日期:2015-05-30 发布日期:2015-03-04
  • 作者简介:杨俊安,教授,博导,研究方向:信号处理、智能计算等研究,E-mail:yangjunan@ustc.edu
  • 基金资助:

    国家自然科学基金(No.60872113);安徽省自然科学基金(No.1208085MF94)资助

Parallel Optimization of Chinese Language Model Based on Recurrent Neural Network

WANG Long1,2, YANG Jun-an1,2, CHEN Lei1,2, LIN Wei3, LIU Hui1,2   

  1. 1. Electronic Engineering Institute, Hefei 230037, China
    2. Key Laboratory of Electronic Restriction, Anhui Province, Hefei 230037, China
    3. Anhui USTC iFlytek Corporation, Hefei 230027, China
  • Received:2014-12-25 Revised:2015-03-04 Online:2015-05-30 Published:2015-03-04

摘要: 计算复杂度高导致循环神经网络语言模型训练效率很低,是影响实际应用的一个
瓶颈. 针对这个问题,提出一种基于批处理(mini-batch) 的并行优化训练算法. 该算法利用
GPU 的强大计算能力来提高网络训练时的矩阵及向量运算速度,优化后的网络能同时并行处
理多个数据流即训练多个句子样本,加速训练过程. 实验表明,优化算法有效提升了RNN 语
言模型训练速率,且模型性能下降极少,并在实际汉语语音识别系统中得到了验证.

关键词: 语音识别, 循环神经网络, 语言模型, 并行优化

Abstract:  High computational complexity leads to low efficiency in training a recurrent
neural network (RNN) language model. This becomes a major bottleneck in practical applications.
To deal with this problem, this paper proposes a parallel optimization algorithm
to speed up matrix and vector operations by taking the advantage of GPU’s computational
capability. The optimized network can handle multiple data streams in parallel and train
several sentence samples simultaneously so that the training process is significantly accelerated.
Experimental results show that the model training of RNN is speeded up effectively
without noticeable sacrifice of model performance. The algorithm is verified in an actual
Chinese speech recognition system.

Key words: speech recognition, recurrent neural network, language model, parallel optimization

中图分类号: