Big Data

Research and Application of Improved CRNN Model in Classification of Alarm Texts

Expand
  • 1. First Institute of telecommunications technology, Shanghai 200032, China;
    2. DS Information Technology Co., Ltd., Shanghai 200032, China

Received date: 2019-09-01

  Online published: 2020-06-11

Abstract

Aiming at classifying the police text descriptions of city’s public security for police stations, this paper builds a text classification of police descriptions based on the existing text classification methods used in other industries. By demonstrating the applicable occasions of common classification networks and their advantages and disadvantages, and combining with the text characteristics of the police case description data, a network structure based on Improved convolutional reccurrent neural network (CRNN) is proposed. The proposed structure provides an optimization key feature extraction process to make up the insufficiency of the existing model in the extraction of short-text feature. Through the comparison test between the proposed model and the existing common classification model, the proposed model not only shows an improved classification accuracy, 2%~3% higher than the existing model, but also provides effective guarantee on the relevance of local features of the data. The model can achieve accurate type classification of police descriptions, thus improving the automation efficiency of the police station.

Cite this article

WANG Mengxuan, ZHANG Sheng, WANG Yue, LEI Ting, DU Wen . Research and Application of Improved CRNN Model in Classification of Alarm Texts[J]. Journal of Applied Sciences, 2020 , 38(3) : 388 -400 . DOI: 10.3969/j.issn.0255-8297.2020.03.005

References

[1] 李荣艳,金鑫,王春辉,等.一种新的中文文本分类算法[J].北京师范大学学报(自然科学版),2006(5):501-505. Li R Y, Jin X, Wang C H, et al. A new algorithm for Chinese text classification[J]. Journal of Beijing Normal University (Natural Science Edition), 2006(5):501-505.(in Chinese)
[2] Peng F, Schuurmans D. Combining naive Bayes and n-gram language models for text classification[C]//European Conference on Information Retrieval. Springer, Berlin, Heidelberg, 2003:335-350.
[3] 翟林,刘亚军.支持向量机的中文文本分类研究[J].计算机与数字工程,2005(3):22-24, 46. Zhai L, Liu Y J. Research on Chinese text classification of support vector machine[J]. Computer and Digital Engineering, 2005(3):22-24, 46.(in Chinese)
[4] 刘月,翟东海,任庆宁.基于注意力CNLSTM模型的新闻文本分类[J].计算机工程,2019,45(7):303-308, 314. Liu Y, Zhai D H, Ren Q N. News text classification based on attentional CNLSTM model[J]. Computer Engineering, 2019, 45(7):303-308, 314.(in Chinese)
[5] Yoon K. Convolutional neural networks for sentence classification[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 2014:1746-1751.
[6] 卢健,马成贤,杨腾飞,等. Text-CRNN+Attention架构下的多类别文本信息分类[J].计算机应用研究,37(6):1-6. Lu J, Ma C X, Yang T F, et al. Multi-category text information classification under TextCRNN+Attention framework[J]. Computer Application Research, 37(6):1-6.(in Chinese)
[7] Chung J, Gulcehre C, Cho K, et al. Supplementary material:gated feedback recurrent neural networks[C]//Processions of International Conference on Machine Learning. 2015:2067-2075.
[8] Chorowski J K, Bahdanau D, Serdyuk D, et al. Attention-based models for speech recognition[C]//Advances in Neural Information Processing Systems, Montreal, Canada, 2015:577-585.
[9] 凡子威,张民,李正华.基于BiLSTM并结合自注意力机制和句法信息的隐式篇章关系分类[J].计算机科学,2019, 46(5):214-220. Fan Z F, Zhang M, Li Z H. Classification of implicit discourse relations based on BiLSTM combined with self-attention mechanism and syntactic information[J]. Computer Science, 2019, 46(5):214-220.(in Chinese)
[10] Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need[C]//Advances in Neural Information Processing Systems, California, USA, 2017:5998-6008.
[11] 王根生,黄学坚.基于Word2vec和改进型TF-IDF的卷积神经网络文本分类模型[J].小型微型计算机系统,2019, 40(5):1120-1126. Wang G S, Huang X J. Text classification model of convolutional neural network based on Word2vec and improved TF-IDF[J]. Miniature Microcomputer System, 2019, 40(5):1120-1126.(in Chinese)
[12] 王吉俐,彭敦陆,陈章,等. AM-CNN:一种基于注意力的卷积神经网络文本分类模型[J].小型微型计算机系统,2019, 40(4):710-714. Wang J L, Peng D L, Chen Z, et al. AM-CNN:a text classification model of attention-based convolutional neural network[J]. Miniature Microcomputer System, 2019, 40(4):710-714.(in Chinese)
[13] Zheng H, Chen M, Liu W, et al. Improving deep neural networks by using sparse dropout strategy[C]//2014 IEEE China Summit & International Conference on Signal and Information Processing (ChinaSIP), Xi'an, China, 2014:21-26.
[14] Wahlbeck K, Tuunainen A, Ahokas A, et al. Dropout rates in randomised antipsychotic drug trials[J]. Psychopharmacology, 2001, 155(3):230-233.
[15] Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate[J/OL]. 3rd International Conference on Learning Representations, ICLR 2015, San Diego, United States. 2016[2019-09-16]. https://arxiv.org/abs/1409.0473
[16] Ioffe S, Szegedy C. Batch normalization:accelerating deep network training by reducing internal covariae shift[C]//International Conference on Machine Learning, Lille, France, 2015:448-456.
[17] Mohammand A H, Alwadán T, Al-Momani O. Arabic text categorization using support vector machine, Naïve Bayes and neural network[J]. GSTF Journal on Computing (JoC), 2016, 5(1):108.
[18] Devlin J, Chang M W, Lee K, et al. Bert:Pre-training of deep bidirectional transformers for language understanding[DB/OL]. 2018[2019-09-16]. https://arxiv.org/abs/1810.04805.
[19] Pappas N, Popescu-Belis A. Multilingual hierarchical attention networks for document classification[DB/OL]. 2017[2019-09-01]. https://arxiv.org/abs/1707.00896.
Outlines

/