[1] 李荣艳,金鑫,王春辉,等.一种新的中文文本分类算法[J].北京师范大学学报(自然科学版),2006(5):501-505. Li R Y, Jin X, Wang C H, et al. A new algorithm for Chinese text classification[J]. Journal of Beijing Normal University (Natural Science Edition), 2006(5):501-505.(in Chinese) [2] Peng F, Schuurmans D. Combining naive Bayes and n-gram language models for text classification[C]//European Conference on Information Retrieval. Springer, Berlin, Heidelberg, 2003:335-350. [3] 翟林,刘亚军.支持向量机的中文文本分类研究[J].计算机与数字工程,2005(3):22-24, 46. Zhai L, Liu Y J. Research on Chinese text classification of support vector machine[J]. Computer and Digital Engineering, 2005(3):22-24, 46.(in Chinese) [4] 刘月,翟东海,任庆宁.基于注意力CNLSTM模型的新闻文本分类[J].计算机工程,2019,45(7):303-308, 314. Liu Y, Zhai D H, Ren Q N. News text classification based on attentional CNLSTM model[J]. Computer Engineering, 2019, 45(7):303-308, 314.(in Chinese) [5] Yoon K. Convolutional neural networks for sentence classification[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 2014:1746-1751. [6] 卢健,马成贤,杨腾飞,等. Text-CRNN+Attention架构下的多类别文本信息分类[J].计算机应用研究,37(6):1-6. Lu J, Ma C X, Yang T F, et al. Multi-category text information classification under TextCRNN+Attention framework[J]. Computer Application Research, 37(6):1-6.(in Chinese) [7] Chung J, Gulcehre C, Cho K, et al. Supplementary material:gated feedback recurrent neural networks[C]//Processions of International Conference on Machine Learning. 2015:2067-2075. [8] Chorowski J K, Bahdanau D, Serdyuk D, et al. Attention-based models for speech recognition[C]//Advances in Neural Information Processing Systems, Montreal, Canada, 2015:577-585. [9] 凡子威,张民,李正华.基于BiLSTM并结合自注意力机制和句法信息的隐式篇章关系分类[J].计算机科学,2019, 46(5):214-220. Fan Z F, Zhang M, Li Z H. Classification of implicit discourse relations based on BiLSTM combined with self-attention mechanism and syntactic information[J]. Computer Science, 2019, 46(5):214-220.(in Chinese) [10] Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need[C]//Advances in Neural Information Processing Systems, California, USA, 2017:5998-6008. [11] 王根生,黄学坚.基于Word2vec和改进型TF-IDF的卷积神经网络文本分类模型[J].小型微型计算机系统,2019, 40(5):1120-1126. Wang G S, Huang X J. Text classification model of convolutional neural network based on Word2vec and improved TF-IDF[J]. Miniature Microcomputer System, 2019, 40(5):1120-1126.(in Chinese) [12] 王吉俐,彭敦陆,陈章,等. AM-CNN:一种基于注意力的卷积神经网络文本分类模型[J].小型微型计算机系统,2019, 40(4):710-714. Wang J L, Peng D L, Chen Z, et al. AM-CNN:a text classification model of attention-based convolutional neural network[J]. Miniature Microcomputer System, 2019, 40(4):710-714.(in Chinese) [13] Zheng H, Chen M, Liu W, et al. Improving deep neural networks by using sparse dropout strategy[C]//2014 IEEE China Summit & International Conference on Signal and Information Processing (ChinaSIP), Xi'an, China, 2014:21-26. [14] Wahlbeck K, Tuunainen A, Ahokas A, et al. Dropout rates in randomised antipsychotic drug trials[J]. Psychopharmacology, 2001, 155(3):230-233. [15] Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate[J/OL]. 3rd International Conference on Learning Representations, ICLR 2015, San Diego, United States. 2016[2019-09-16]. https://arxiv.org/abs/1409.0473 [16] Ioffe S, Szegedy C. Batch normalization:accelerating deep network training by reducing internal covariae shift[C]//International Conference on Machine Learning, Lille, France, 2015:448-456. [17] Mohammand A H, Alwadán T, Al-Momani O. Arabic text categorization using support vector machine, Naïve Bayes and neural network[J]. GSTF Journal on Computing (JoC), 2016, 5(1):108. [18] Devlin J, Chang M W, Lee K, et al. Bert:Pre-training of deep bidirectional transformers for language understanding[DB/OL]. 2018[2019-09-16]. https://arxiv.org/abs/1810.04805. [19] Pappas N, Popescu-Belis A. Multilingual hierarchical attention networks for document classification[DB/OL]. 2017[2019-09-01]. https://arxiv.org/abs/1707.00896. |