基于深度神经网络的单词预测系统设计与实现
Design and Implementation of Word Prediction System Based on Deep Neural Network
DOI: 10.12677/CSA.2022.1212286, PDF,   
作者: 王 昕:同济大学电子与信息工程学院,上海
关键词: 单词预测深度神经网络模型系统Text Prediction Deep Neural Network Models System
摘要: 文本生成与预测是自然语言处理中一个重要的研究领域,具有广阔的应用前景,例如通过输入法或者检索框打字时预测下一个单词或者文字。然后人们的喜好和习惯不尽相同,传统的预测方法难以有很好的预测效果。而随着深度神经网络的发展与应用,利用深度神经网络模型的文本预测系统识别准确率和速度也极大地提高。本文训练并评估了最为流行的深度神经网络预测模型,并设计了一个单词预测系统,使用前后端分离技术,前端是一个可视化网页界面,后端采用多个深度学习模型,方便评估模型效果。
Abstract: Text generation and prediction is an important research field in natural language processing and has broad application prospects, such as predicting the next word or text when typing through an input method or a search box. However, people’s preferences and habits are not the same, and traditional forecasting methods are difficult to have a good forecasting effect. With the development and application of deep neural networks, the recognition accuracy and speed of text prediction systems using deep neural network models have also been greatly improved. This paper trains and evaluates the most popular deep neural network prediction model, and designs a word prediction system, using front-end and back-end separation technology. The front-end is a visual web interface, and the back-end uses multiple deep learning models to facilitate the evaluation of model effects.
文章引用:王昕. 基于深度神经网络的单词预测系统设计与实现[J]. 计算机科学与应用, 2022, 12(12): 2813-2824. https://doi.org/10.12677/CSA.2022.1212286

参考文献

[1] Paperno, D., Kruszewski, G., Lazaridou, A., Pham, Q.N., Bernardi, R., Pezzelle, S., Baroni, M., Boleda, G. and Fernández, R. (2016) The LAMBADA Dataset: Word Prediction Requiring a Broad Discourse Context. In: Pro-ceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Berlin, 1525-1534. [Google Scholar] [CrossRef
[2] Cho, K., van Merrienboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H. and Bengio, Y. (2014) Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics, Doha, 1724-1734. [Google Scholar] [CrossRef
[3] Goldberg, Y. (2016) A Primer on Neural Network Models for Natural Language Processing. Journal of Artificial Intelligence Research, 57. [Google Scholar] [CrossRef
[4] Irie, K., Tüske, Z., Alkhouli, T., Schlüter, R. and Ney, H. (2016) LSTM, GRU, Highway and a Bit of Attention: An Empirical Overview for Language Modeling in Speech Recognition. 3519-3523. [Google Scholar] [CrossRef
[5] Chung, J., Gulcehre, C., Cho, K. and Bengio, Y. (2014) Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling.
[6] Bai, S., Kolter, J.Z. and Koltun, V. (2018) An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Mod-eling.
[7] van den Oord, A., Dieleman, S., Zen, H., Simonyan, K., Vinyals, O., Graves, A., Kalchbrenner, N., Senior, A. and Kavukcuoglu, K. (2016) WaveNet: A Generative Model for Raw Audio.
[8] He, K.M., Zhang, X.Y., Ren, S.Q. and Sun, J. (2015) Deep Residual Learning for Image Recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, 27-30 June 2016, 770-778. [Google Scholar] [CrossRef